var/home/core/zuul-output/0000755000175000017500000000000015156320275014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015156324423015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000115622515156324352020270 0ustar corecore꨹ikubelet.log]mo8~/;6M:`4޶-67̯CJN6S9>Dض?rsG&}WA4ItnUË7˫ɬau#v7\漙NscWvkVU-<-'ʲaSe# 2(Ɋɜr)kx+&YDdgd-ؤY |B,' OBpЭy'C7.Ɍɪi*Bf2U SWe( Gy6/Ѩ'Dς5BOU(5Mdmm8&_TY:\4 ȳl:,\` aMm5^2oa.]Dt/;kѬ=8N%3f[\l7 Βsw* 04Sf}ɢwVZE=@1XQ('>Z/Oy첷Y!W:Xs+Ηw09\.3X܁_ZHnّ恿v[)RQ ojQE}D+ިO1SsUKh <W|!JpHΖJ3/}5q֥"ܩ m;(ZCwR=,&߾p#w)KYUkсN "V6}o1F'h<7vQ:ғhtӹG +D9]U8&1f"kG?2G9lxSzyS>=f("P$6m-IAMct`49"3m`Z t?fe&CETf%ޤ[mñNk?&uW5s1t9{;(RiNV5ݏ JwapPJ9-W,xHQ_8 Gml0kHwXUƢ%hCuo5dD|4HIPc0) [+}<ekd}x鿕,qxWB\:1E8\>W9Nm^ˇi-cdz# e@߽O<:'SnrGWT7!02)]hTZ'](竎YЇ4  A \tE.NebG"7Hs&W!%-47#|k=dRGm*kp6,,uwy\†rVTv_a9$=ls&[(8iPKX|y[+ $}u]tl`Ge|ڼ9BS)Ov$C k4JXƮ;iCL7]֠拉C] u~#n#m͝fk##S@# ;m±>iF!{C/~#uW =n3c}vIJ5+d)th\GLqC mDǀ֞G U[6! 'PR$C;NS΢_[^jG <2sX8+*~ŹPr^CkWYb>Z!TӼ7Pze psxOLÁ;bomo!`^بPz%U}~G ȑ EŮx`i}2` ]57'`nz!@MB\7 :+\9 J`t 9w~3j!OU]iV:҇Nt >isw [{V ;)FV je6O*La7r.p#/ [=f|8T&wYoV@b$Q wU,ͭ# 68pZ:$*xb*;.Vmg I8 z`,$HpɉwDV utK/hc#nMcpGd#i2,pC)*+WټU[oh  @*pQN 8 ¢4PJq:GJ>Kز,#?s,0xkO2P7e}n\!Mݏ08C|Y~3xd 5&wi·{ч J`T{?b:xfx 30q jmItlYɏ~LWuO栶~GEvTx{`!Tj#1D=1FUmYb[$pT,7ȩ$IpC:U P]X-`knڬ3 PtU1_Cůԟ.'.x׾~ę14r㙟 /ÄGnSk=n ?1a!3DݝA P _^hh}P^xlbp}lС]4W`@t(U*[&:וC,ۭYŋP6Ga$KB{(4qwTgmUYAErPm~{B:8vɐd^+^ŁE`ibI+XgcTumZ1Y>ѽ-h: ǏpAF[zRo6>ӻvŎ2u[2V,7ww}XTWJ.׆#21V=Xg6xsN0*6q aL6-M; ߻:A6?:ؤ(pCeZҌ|ElFbJ]K],zȸZ|H/~r_ĭ}y:#OnGLb.+=^+ˆPu[b$XD SvҸv0wnV}(sSop܆t7ܬ$=?@mr'A{; }Za,#Y$0$+gݵ>!ѡ-+~keTJryk@Y~VN?^RW݅%qf^C wrB@vΆmNedGϨW Oq܆*Pznpry@izw1g^R@GܑL)DhFJH^@4xR"ѨneFtu|V>Q@iw#x!dۑOME_PGk zYn d^5{sGt_,A4x6jP2e@& {f@a[m Tb2$@.p8i*H:JDbu@'Lf^:糵WKƀ? (dTzQl'lldݧ>c Oh*K%s~i' ۍ>=Mַ&i "@V' Gc=[ qD0|j |R` :`AnZ#* *u7?o/D;f`#•Ds)d77r:J J~; GDd H Oړ18O2NұujR JD |/"Bww]NĤ`, F6S~/8&Ć @OA MSHTk{R%.y\q#?Diے"}PЙC":m\Юs"S8(Baj|ƑBýwC}P"4i6mi,ɉyodqN4,@_pf=(UTհmҾ4µcIĵ⭙{0*RY&'cM+ 0u :`M=,Uf#]^:\0~OQfw1eLK΢V|5cb0.09's%hm"go^_y,5Y1,W$_?NDa>Za2'*/#C۞6i ٚ^&ѩ8ήDN!Mb]_^xSL_e])=e+R>,ࠩ_!+_Z7x@gMjUj *I|^C >'@QuL/U:̧h"|YÁs}o[tJ<'!jcݫ?O}љã~ѯSNQz[! ^0 QkW%*J6Y"r]`tM"=msDPEdz Xr5P"[ܹo-8 쵇Bx5j3@c`kDXd x"edo6#ť DO|/ BK*J`%,2<}Hz`4o O.lYF,f"|K\Kd-S03 }a 2# "~7v" c-x|{LA >plfF'(ZX;" JU@f +`OV`q ,&bv(0GAvDH DCh=v'}4 BJ툆P| bG^Xb8"BFQ:sDX]X"|OmEPmF*%B ǔD { .c}رo7c<oQ2Zx"|\oUTDdQ@G#"Tmc `=/=j5& Z=7Y~| FX #OAL=x ,Qr? DGxhG}ٹ(~b<re|@`y cicltgO.*Aǹ{tR ")e- ׵K\oոBouǼ^u }x ? KǸTYF-b 4" ¨@ӆed. l X!{ds =^.go$ u܀Z`pַ1yU'(t-"ćSLRRԤ>!#:divwS^a֏p08]cXE/W%՜#7ɿ7;Kt!] |pTbxߢ?-RUģ^؋bR|pqyq~7J>鄔v(DT TjD.|t\R_nd-NVBD@}q0̒"Z1i28kh~zgOL!۴*{7MfD#Jɕe9qg˙ WSu(eY(kNwrщQrr&#Cs/s9@qx&8S%q|=YaI4&G׺\Y'N3Pf|,ZۤVo} p6ח pRSHNd嶠2'\oi2ss0kքucr <s9F{8 drܕϪ~]}'*y:#>sT#i_pT:2h ݕ=d[ Nt<LPM\R_ULrrxڠAA^yS`MLW*z`s5j5ܪ%hjz ̮a`1^VUˤJRZ|XwUyeMYn2-'3F9kC8Y`ںQ(* k)?~Q*V h2 y9:)fe8>\N[PzxrL.&G ~zySÙ:ә[eԶ@(YRڢD X>H<[O)Vؘ Jeuê}F;lǣ&Y1㑈Wfyf9_.e5EQ/);Db(Q˲pso5 EE0N}h}{ѾGgY[СŶZX~I8*[J/~GYxba*=R{J43wXWuO3Qx!9t\z'y2Kb(΁4ΛNs;E?I&=WЍ|$?)$ٻL}/ ^TM~cؘ7e:)f2`)@ۀv(pSA *vW 0'[~KJ#Խ'. ˘TwXݝu}`w Cރzz;N0R{R@J#Կ';Nh0B{@([%@(۝P0B= e;@h; Ih*F=螄F;W ;w'?P~OBUB #TܓP%]R'S)IhNA_H(Ȧ[uEqD):۵|Z=M*.+Gbx-r]/Ŝ%v@*Bje'h$C+yռIRo [vv,*O}8c"7iEϢ<yQ<)yn9H*H!)ʼ;#ĥۣ1NŠS͗Ds_ FtnC҆=Qh24HuFPDY?ՍU› iPS7X&VxwxǶH9=OdԀV?qLAى'y{2؃$ѓ&R~~N'x՗[#㟍!3ΛQo- opLdkH2 []hUhJX@}yn(e#m^;zX/8pdYb9;3 Bқ2?dS1uOg&7fYJsiUwk0*+ hioI;X@En A~n7 pLF.Q]1y_Ծd8A\75r}=Ț,!,AdH11Er4bI'-M/ۮ6!fOPM}6`ywPlͥ* 3i\|QߏxVsiKMdm1ӞT^" tɂ ւJ򔙼Ꮁ6AO6]f^WT8f8tBtGaƱˋ|&5=Xݸ #͙Mh q=0;EDNi-/e(FR>j+W6ԒBN8$a0!YU?\7y"K#]qmp;ja!{ɫ|WZ4O ލ\<}& e>ƣx({99]]vfڪ<(C>%rPV!Y>p0dTc}F cx?UͯO0>nx`!TtV'[ިҎPMۄC-I]ƺ&n`7rZV1l_`映1DOx*wͼu[c_,Ipܹ *zetg=_]&e=ڿ(*4<>-N#x~_bn[Y wY7n:bć ~hGf?M`7$׹a5]% LZߪ*۪$Ze(XY såҕtɎVzT 3`7M !u"Cu׫YiHeY{k7} tTqH<uZpƲzzT:jaļ6Ƣ<' R'`NC$i9c$p-5DW>ehgĀsՖn8L)yGE/)KEkyS:[ ArE Xwtm]Z@nû λRlʋ[HW~uc QoT €:CuNCv"i(sC`X5y&W*)qEfX&51ױ nfQ U&P#>rj|~YfA+CWi28 X4)otIobZ|o[ekpϠ~sq3Miw3M yBM^־QgpqA!MX9F; шfpE0ϛ<S`J;_^6d ?WPo!h!'pjUpeAv|缙T<|z:PR.pg3 QAU2Hڨ7YA54 kV ~8r &KN\`b0ӹR>gr'zQ:V9͠tL*bj+j]kocGr+? Z!Adw ; ~C2IiT_pvv ,<],896|rKg܋l)SLmľ]فwƶrl^&\ȴ^.+MY:t^Ccu0k j֒tbn^,Ǵ \7;߆;y0~?G1}_O:SN2΀~"_ l7 g1͛3vÂ]18L_WXd51ǻwv_ͻ~Pj7_}qoZTs;fo?g븛mŷ#k5}[tq?`qr3oۑ;[in<`̓8_AȷSOD~=Q׊$9~b?lni4D[+<4 ;wȜrG\]?8osSƏW[svо-OU_]mnM}X.Ϻy;9׀D[Qe'np'Pv>Ó>;^wZ0Sz>{=Qo&bNy{Zn sn@Fn@OuE}[+a97-v6r~i+)&~Y,'&,13AWW0<~Wa 5WgܦB6\l謧e'7鎃܁ *mgܰ^nw֤{8nwVwnwU6dq?%`z]}ι#l=SIa|!Lskg֌Wmλ7=6UR_AH ͳD~uuzd'=VfÆO \/W84C5׼ڬ^OSw@S2},Bɿ9eG(&4:NtZA:^gWf1 L}txî.&t\W-oPhӯct՚>o -ioiKa90+,7;1;Җ>UtoZ -9oMC}ppL;T6ӏ]S`Yk೎+ KMJx +αlk$80^IܙwhQ0>G_|L{QPXFf^32TI'8ݸ89jyZAA*& dä) "ll[ rB,&[SdR,Qd!Rv?/Ii]J9QzPت-"9g0:#R9R"4(`{{ guic1CE z(ip!" Ҿi@`U'r!,֒n. VZmE%w"i܁?;u Z W j)bE.lh-i"?l.88s8hnb=:ugh"LNXJw2w ">B4AlFf-^Ft զpQcT0?ڎ,5Éf`7yN2^7mѤ*ˤ1v1/p.wԏ~(&d@KIP[u"{MB"i6RJ%wlV-]^ [?y5(.& ]]`tqVcY%j|11ĢVFRbfQq̱/N٢7΋+JhQ)=<Y)e;^H, Mh$8.kigKԅJ,=2aiElfN %gzvziE=H{vZ97Jr! ?fEեf)u!׈:4D=xgӊ`}z18 z3Ot e8Z퐣'}ED*&tէG7?7=h(Whmov a\].6.0 ?PB`h ƴ^""q#S$u"f9#ǏRr;,N:XL6 9xH_m#pU0>xPYfuS%I_L4Yꢊ3y2c\vq sk8z'tđ͇>88̋9e3箾.wz{7тܰ=,E#(zF㧸'gۗw rIU2XV!!ǧgͨ& iX1#XSRN,$.kUe XBc!̈ F)]a 97ŰjJeMBdٕ$+:Aҩ.8t.rkB< nQN༒RA`{۸vۢJHAQn1l۽%Q*dCIe;ӑXNEy%ZDJGHAi7k312 lj>X%ة}0~Gӡ۾Ń3bVNoɬD2ۼ/v1he ӥpEզC|Jg^!ԛܸ&[^/{J*F=7:Vp u힍=uFhǎ 7=<\cJTv1ADoeX> $.#ur~y[D4]jEbHu PW1 _(IDW +\i'gWS RH 9{|NQ#6O;TՊHuoLbv\HJoBl7&FRQviS/o[^Uvt9Mv!QRVu9MZ 3{[ǹ\zQ+8*r.܂^\n]R֝KM_ޱ l% f:Ơa62 M{ukz[vspԾrf3jϦvdlPd \ICk44:"XbE-Y8GC@ ճj,;$޷7gQ:2o:ϼj6gVO`Z' S7$6'ײf@Q:;g}|2",n@Y4y9 P11߲k綅,JېS.&&ރ,'UD/'8d#xiG p@u hKQlb bI'z8Ӱ=vUhdFkX)@jŔM+ؑ$ߏ;l06r&$k2鏤S| ](f: yF9tC ֧e:Owi : c$0 1kP$8fHiH@5TAK}z9z8kF](M57Enc( 8vܪi PsSvܴE A!%{ut +J+*ի4!SxWߞ:`[.+Mov :.q[季֒Cگ ߋK]cؼ]CjY[Aw+ Qʄ!BE s5q;EN5'@4+GF xL@'V1,1LM7tibVpW58FReTǾD` ŠQ`G 18 NU]i7!ѱشERWS${ޕ[Atm(SFqd@fIS.`ݑ`>0j(o;3Ҋ0?vt ;V ݮ.uys;nkq|8~ͨq\rdF,U1ZE?VpS/ 33'[rJ/U{ɁNvc/nvh|co|"p}00 `"ǫ.#m69!q.,ذ㴭Sϩ~j{uSեࢺl ;VA< !-Ѷ,Q /}9eͦ*r#?]+\ԦhO=ΎMNbzc0,4+2OIpcJ.#m^ 3,.H钸2L^:.д^J#D)ʴ=dl[{5V4[Qu6b)u^i*]]Db:җ0Eݵ\־,_R*g&kyTH{#!ey ߐoڣʥGIkUDp`: 0oBx(ˆvC2rB*s8!6ѡjG.{*.|.zb&8P+M}hoZRrs\nGqa$0 I(ԵQFkA>2,VJeF^gD#2t r"V\BRdU,2zxQ v(V*K\US,U^]yF3qWkBw@W,oW]2ɴLRݱ4pr9?UlRH8hߚdLZщQh(b-Y1oUY&oh靋*'W[OeyIrp c8Tuy_WlW-훒y''%< ƂCqQKmx lY4nwo lĩ,-d(=t'$՜Ko>Y RgaLN')T}\$*^>rRV7O-vlZjg%G8m-(EA8_:,գ_s&3+v<]8g Y4e8-ybJQ|ß$&h~^9_<~{ntBq`iPxlůe(`3!-W q 0V3T0MЦ*m-D͍t#8a`ѬynwP~uq$id(/l”>&oCߘ7g˂mrLyZW[ЖQQ ' @y{)%)oio졭X d]娠H땨꓇/*8#GB0/nOMAI&0-zn0vr"/'_; C?xnPa ٶ9dD'ގ.`bEhAqye7'Ŏ&+RDx+i!^$BqS3`+ne}ↀFeǠa0aĽ~J>n`6Ӄ~ky_l% Fhѹ;6|2;bfwѤXwc HLud>\euⱧOb}oMsf=<PF R{T|{EB-?"ToN`(9i?lo2vR%by>])$VW/&YȣٕÇR(}R@0RC8lMPy@AW*P}QZdDx_HW+|{^DDԵ5aX@{ "FpH6LDs%Wݛ2quCPaVUGd3{/yq3 Yan6o4ИF:!$Dkɐ'R97P/}ζ0][(@(<wucj;X[Ѿ` ڸ}Nљaw^:|h3+v(PJGyҊ.][~=<^ssgp=6eXk83Cߏxl5[re{zVcoW((gJ&49nɕy&kMr 喬e 9Ly#| .N֖+eKq'@i+]^hVZ+beVNmo7m0xhڷ-z`,p[w-Lw}ߎj('ji/R1:Hoڿ%6nJtPQ܉"ϭ((ޫܔ(:B,yWvޯcO@9z97w^اٍw |<)wgEY}KW`/Up\;V";jaM@+s}e/j E6])[NuinRJ*)C悷ɫcval4~h=okg^ܮտe *n2ǻճBygnjΐW`I>L#s3X&/a m RXA`őIP.Brq(A4%( 9&Z+ov% % Lqi+IDžǩϼ:>2ɣ~]syb+ni,o?OiOl-Ӂ7i\gwU e,^}6@@,36go'ѵbCޓO"ZEM>Em Tp}zwY7O?&;Hj!x-]%6OY]RGHŁ3LBqDPqHcd"%4ՁQLdd['ΒI:g@9÷u"Iih#P(W L-eT 8T@ D3YcMԈ3>ɜ !;بsBo,140bZ WAD#D C u挦jMt v?7glG EJ)lLcdš6Z@(X1#X+DhwAdFÇ?Yphz*Y~ B+ uڼp1XәO{TOR3?#3҉ի~5LG!"rPQ3ez8p>e|m{`AGCS/'G6&~H 뱠]ǘ*Vobyc;a}ST"WppF6Ӌp~IJʺPMѬrʌA,?޲7 Wnd&f̓3**%|OLPzaVc|G?N&S?7gն\ vແUoj2'y (T@2D.InK\m7 UNDn!H*Tۼf.P535!1U4E\l_5D^,7hgc+/noST  `s4].ObWi|l#>o:w!T^5(+@@:e#e f> $="0f &@\ Ca_#Bg2MOuNB>g*fKܢg;N;e8(F 3sQ o6Tia)'+N>Ngb|_MW @s^R^^ژK۵ۥ,Cms: F\͎d#+?bD ,q\`;Ik,)QKRU W=3$IJÑ(4vWUWl~K0AVM\q~8f-a[&pEcUFzLr)*c)/na6"~ˇC:C(k1^ Q|T&޿%p9i),L 3x 5aU%BH\4;48\Ni|C-cH;C{=xδh!au< MhsL&[O R gเ])ۄDRd\@G@0w 򬠜]jEN;2cmgĎQ{2-V4cǭW團FFMASrfޚL*wa8h;2)F[摒6h.c8no:VPfp(b3H8 p!lp,X6]Rl:\Bp)C:b;.P=+CL/pCt&&N0`sãJOc@P}W`XbMZR;.Ylԝ)RkzV6E"X["ewU-ꭇ\),ٷ#}OqO!7ye~\ Z9:|G럔 \4qjrTagL2 :J|uG' g2W~3O$UEƆ`1E N+ M2ST}8Ԙ8 ɞ A"2%3 5s,@ $yBV2gfFk5lh Z"ٺF0-{i2/| ~E$nP(V2(T&plQ6V JFk8]@:9րj6c}2n1oeri!TDM5VyI'➑Ro&P Ev1PtOӵ*UgQNpp F#"E<Ҝch 3pFdLnp?,m#`5]gbN M Z'2##_ Tk(>"->m0M*aɀE@{MB1X n ^2&Ix"h.66/%37^!`&GA>^oqߪ>y„1DJ\JV)LH;[.)kL`9-ba2ƹ͝6"'%CcA׷;[ %4xIdr}j"ܖ%Xc$kXrKa0 +ZzY->֙V^q2ctʚ'b(2) m\kìZb.fE(0W7 ނBZ74CQ.R͔-4%RTǴ*ir; uLxǴɩniK LЅkCY1ϋ!N?5g<b8EM*>YmD9cN)zH(L$EݹU+N㙡Lf uqKl ށGJ9%<*ObdFL4QIQLe 3Ah@ɔtBYʅ^"츢R D* i!4R͜Ua1Q8m[I-&)-e–~ t@[ZJ"ssdje0e:`^pK#wk9g*)CRoNzs*;\H}Mj Y,ݣ"I>]Kk *P( pe<6&(.5`ю3NR,2{` D)&*~lw+dK8z3dRDu!^6aO ޅf42+kO; ) 7Ey$ ŖF0Q#X}4rU;@Ij: a)mCRXwl>Yj_PKrό;ʯCCxxZC u}W#!%dn+V+t.c(XkyzYlUL\LF"30<6Zޫsc}wuw8w( v0ȫOǠ+fy?[ N&8 C$9JMe>?Qn q/σ1h]:P6o&\X4J'< FXSȤ~;`r4rnDjk\qRaBӓ?@ҪdrIe904Rѕij Z.iC۷RGmjg˱.C+2^S8DQY1x%$p[)dbf?v'P؝F.iyׄH-H/~%i=Mxa\F"ޓ;qr"j2$by\T)i>Att5^7\^m.Vlzw0T7\h!]px%2UcDF.k9Z'@)k_w=*l66Mf T0K#XJD.C-T&fDj5Ƨռŋ E$Zk$2T'_&Dvi䪖whTMĈ2ykV?(B*٤~rwXx~a>^ydjxH$̸#i_’!$v{?狿^)J-QĽX"}䜍z!//F9C`0_{yK!4ޒKJe\^4bW(/[/__C_Ls:0~VБxR+^KI1,Pl \X.-оLsdA/`G9#37yѼpKsŔ'.ϱ?\`dX]DQ/NA?~yW_/.SEY0Wk|G~Z8mg!fԧM-,.Fg_fH? ;zgVFP?^^OQ?b*b t3/b!ףD+P%"˗߃)\ 8W`#|@AxK39-wa6_( x>6\OsAZ+a ȇ+$?ƀRvFJ">^R3\Qi~HLQЅg}Zx)^} '?9 Vfrgvv}~=`T/Gjzu{90)щH^:[ښs#6l  f<D^#K19NINZlLZ@roy+P8'Clx{6M?|ԏմ\ד{1~< Õb ! 8|F.x \tuxu?[Gl|Qkeqyp~}n<4(L6J'EGx bYܱZ 7M91l{хu[K8/17{"^|0|X3!e_ (^(#(TrV' ꄤ g[YM8bC{V}o<,/cnЪ"9:fSFOw&C 5y~VVY^3PLʁGɵZKmWGvtlcGfae[ڵ\uĴxS{sFZ(#`!r'K}9X*w` /1SQ0դLs"\ܫ"R(Ť5lC4J 2E8C1(Ufmũr\7ԡ~tmR5h$z@2o b&ok+]=ʠdP8N&F4Evv.ykjՊ}4 i%=9;I8=DߍCsk{;1g!4VXmGHߛ88/"_: gb LZ?|b#y_ .|-+dTC%Fl>w\I@.!5Ѩ7yaWT},@oŋ5h1NXiS["$RX@?pΡ$2SԽ{P` ָv'8PɗfI?T>Z h+=f<Ȧl(`MX!&bK7#"c_}Ш$ R^m:ıw!m8!X{1T Yjqi2Jpe]k&skkjVrȨ&{iԳ˰<_D|c58.jvånqv3JsqH>&; `k%,Z;\n^SA'_9=ITcubsq1>%\B>ǭAxI]Ykqh:r B˝?1I|Y'setRK 2:O`DxqaRs~q!r>q!Y+xւ%g^S.E%6r$#Lԭ/뚬<)+lL@@eQR̺hV;ZaJY|cLh&4lW_P( 6F/f( ͵UjYw}C^&67A-.G|S`O_?g:V/_@'^(][^3we791LYC}E{PmI˄f0[k]X.zvXhRXV .hBQ}Ja㻛κ{v~ެJMad{XMq1N-Ok䝥dݶ+.=CzȺ0 FG?j/I bxݩ_T^>ƨ>EtjzU /9=ڟ{ 8Vꇊ']Ec~;=Z,ʴ7W?H3iDK#C-f %-փ $e%;6! (d^nc[N}5)"&!Ɉ H:IUfHG)| 5A# (Sqr oNX7._~><7$FnXGH:Z#AWԛFdÑZWpsHj;=7V WϾ]jf.gQ2 AlǃF[IBDp/UpDi`N*sRX6֢f #mШہ/URځW] dsIf3LӑumfiH!Ӝ.& "H휎H<#moGxK 趼˗*"eۛmm:d *%ٳͶw1Y0E2&$Rc(l{l{A= U[YTlͶw6dj6 zfۻ@ˠHtJ"u <ʶ7ޥ^ pb0+(&TOͮwW̠fHatdJN{]ovsd;!;כHygz]u@]Ftj ۺ[T<ٛmBm72%R^eۛmmow.)%dDRl{]@Ʋh vJ@˶7^unSitdYζ7^N HcJ"uFl{]2h@a&`ma/URmo8E8$SXOG_ywQ{æZoj,<6gmB@;IzK֣ C4+UƇWm%9QY4&7+./~#l^"~~;Wξ1>şg6MwI|~]]|3\B|ն&(Z#$h`v ڡ5\k.:*a Յܐf_n:Si jonTQyGq$8L4p\Ul a*W+6‹^<הmtGPliʦ;ъ EP74  5 UeU۪E٨]g+\8Z~ ǧ YNAR X0Pe!ϖdPE#S}Qmjj;hVUXƦRЪ'gZגc01 ¬kJaǦy\k]Vnz{+!]{AuRC.5e( \D,ZWU8ODn"ݓpfP 0W!?(NHSs ,Q2͘: ~-N]V Ān_dFq("C v^ќLQ9'3'`C6 H8c hd@4U@DU7m.] NGh{\J!8H%IGL:53"0l̤L4nfeCXdA9J(/OՈXtjeV\aH+״m_Q7PbUUa*іЖJRXrZEAlIPb}ArKC)}`Gfi(V ٢(IFyknV5Evx>JZuoN˼gvqvK4`ȌI2h2Z.Ok{$m0^מL bAw1`<ؼ`JoZJ3`'?R=7_x( Lؾ* t2—@h9|4Z>p6+QrdL8I1vAҡ]8 :W+h\1MM&`rp:S2F" X:O~]>!?wu ߿|&+_߿z)>_}S^?XX_hyn~nXUUUǦ"tjj΋s~ҬUqu%cu~{F3f:i=ԡ4r-ܮ_߇3yk0 =ye:bOjPu]s:orgތf#-iLj/ݬttGx @޾gt.+~/w!ᗟ޿K⩱n}YmD(MYaS5d**+f]Xb$)9_6|?-='庶^{c{Ņ_ᅥI+~7ՇwTNb}vm8qS\~ kQ}[ ,']iu,Trw=o1_/},sT{$t}x˷+]gZk"c YFbq6TLdReh{(k+0”0ϑ ʾeT|s0/YF3+e$* R6NcAѹG˹h!Sɶ_ X4xX`iu`>;*H2!Cr-n+kqϯEUyׯYaPgʏƃnLA$Tؿ'6M^~d܇m|X 9֘HGV|pp9W{[[WS]KK>DIi3mV`n`&V<7O VMT3$Bh7LW`]"yɁ}fCãUu_;P 1$h$@\c>e8~W/AW/Ā GcT5'"H2^\hUF˹!r%] WO;=@Id. ZT:O<ON@=*g?ri }0XDj4H24ݺr.[~'1BS*Sz[Tfjtc0‘.Y~ Gg y W0&"H:,דvSUܱuCCO(Cƚ^p8OF!H,tDdHeRlANLCbzBOVcfX" T6MxBRkɱw\I[b 0eg14ÚHF?~ZgGhu?vx!baTT\. o?Gujdbblyw8V }#p\ũqAO`k7pKKp ]?" 8˜o+l7ƙaRIqH'7懄]Alr~T ^0":CtG*luZZoZ [#O`2$ju~M|1|f:Y+lGw1FV;ySa q)s}yπrlrPLyu,Mgi!UI/1M'|>ccmSF"5V{T^V5(3(J/< ox |a^҅~!/J{A3=8{oI4)-[鍵%?OkabwKNh&`|u? JsH4˓OvE.. hCx*Ku㤗>ҙm, 1GAh@ir\vxOQ%ߏ`5{cE=}3eBڠ-|/)30*~ۆGGs"^koe${Ws3 bp60xbу#a_tԖXCv}ɲ%.lU!ft{*3 2ƈגa2)ToMI-ƳV>e7HZz;p*jTORSH0*[[xKʢ'F\+}]| Z{hd.Z~ e:8JdLcnM0W⤫qDT3Q*3D) b[P( Z.JlepiJb$똦 SY: #D~=HK$^>{!PPg嶎( H:h$=LƜ52E\]Hj*ev$i(BN(D- "@!B^]E>%RIu&I'%_] wv͠аAOW03aXz??i\JAl- K9ªc+ TFr$Cs.f*͔I_o%x]pHQN.].0l1c۫9 b+N BUL.+矇}>0wrLr$3+7f"͂403>̩5"hl Vu,Ы5o^hblvB5QGQsbg#L`cT0?1_e3Pf6 YPIM@ɽ;Sk$TĔTϒbA˵iVHye|[ (52v܌[PX S>&TG\+&` i:O&(jHVQ064e% ) :Xo۹x 0 fsڴ;ʶ6rMHj(L8>V,56ʺeb0̮ƾZ<(ŧI_8L f{;mK 4o RFz*x7K{_-ru#'kY 𖷭J39ꙅЃ'm/KkKͦk tqUǂ)~8+yhHUJPQhʡ+뚖ˆ^}.rӗ2x!8}EH4Fd  W/T69K3~Xu-fz&v2aXQ፛Kօ`t]0N||o[rrUm y??Ո:*>ϒYf7NDPӏ!=Ԯ55Q/&DKqƹ!5²טƔdx` DpiX;{Vy)1[ހhM^;gw=าּCWjʡH`BZ}Unph)JdvZ*A62( r&zV$YRtGC$*N &7uOT:¤UC-n$g7߿mi5՘]e8YI2R%R G].Zta~Z "l8պ)YbfL/-n>l@ᢧg3֏CWg'n{mg`x2a4~S0=31̛'~5=UᏳq?ξl \!-0az]ZR&A~@ߙy`TkVA/4 \-CWU>(ڑh~RtEc|O%k$&JzKH`\vKP\NN%k&-x1/&Ұf(npq;j ;boxeeY9Ʊ`P|? \? ĎqTQ)=Ŏ'3-011x@]G?Ŏk.;c kۏȎm"&FGG66۝3{s# Ȑ"f(8̯p ŰvycmD3@}vQ|K\m g#hLxG_!fE!'bOZ@txKVvO]0()Nt (lxA>=afpo @ZƸ{o[w>iymsCx> Ij+Ij~wfH-}*)(1\oZf^/{q7IdJ !y s>R,1%Z.(ㄦR t̒H,4!]*XUqc ȏ@L8F'6XQNڀ9Q gGuW`$a&{!/t HvFЭ/ M9W_~(5Jv) j/i F\s&sNsVqh^vI xnlP~,,7ݙN:ᔥlsoo_&d'/j>?ft{9#QFbjy򛈍ֈb4R KOv yG8$䖓6TY쓠$> e&AA4PwQs(H*Ƴ.~IP{ԊW) qAdH1ԎY"69?pieFeM,OzWl1!]Bk.f1Rv ͇?,m~,:fàXhP%DfRXK/tќY?,KřR$$">)p G2}/Ds)QR C`YO<—n1{qQQq{`%c{.XCDFbjOqh"0Afo -DVNqh2 TV.xz Rz8xå{,D5 ڇX- ྊ/4ӡk >jLj+):F#Sis L̖k|}r1)'P`NOn&§'7 v0(s#9>Bp~}Pppjg(1pp}AKXݰZJҩsm2f,Idwla%4q~ovЩߦ5fGcYL@Px Vy*i~Lj1d^o;>e޽_Y \ myL>}4M-d{Ɍgo`$wL%q$`$dtINF`l\ɚ>Gs \:)޴iu 3f;mlu(S{t}5`hr6s{ÑIԬ aO[S]ۑ}\c"VtBys\Q0ݳ_Z}`0GµY%:%cxۭ4邯t-Ozs "tvե8}8 zS zߜ M{F;]ޟv|$jExIuԫܾ͗i&uj3  4|>hVi~͆Nw]z۝:'+D-*Ћɹh~{wga#2mPȫY;GP v NȚwPd_~ҙ#]W+] ;fRjڿ~JgG9ݎqtX}>n&(ܟN7cשp988;PǼ7w_;aۨ]H%lV B񶧦oFۄ񁪺^ʏ)|XX~FmTT1.ޢxü[?Yf8H5'͇37JyOgNe5#iw볮4}Tۛ9VG-n:KH*ptFZaph{8G[Tt*}b9߾]o!Q :F"h7"|Ul؄תI[ꤰ$krfM42Q[݈ӫǐZo^m|tzQOpQrVn/ZUߕs gOud jxt2U\t`Z }(#dF;dԈ]T^+Ubo9O=Ks `fjC>uFfc6rй+e.9Ur,ZrjwTM ڨXbOuy4kSSNqM8 H/ے4EڪۚIy], ahZP #HC }tof[9egaDpF#$&ʍZL ѣ8qD>L[k)9YI y8jCTmY;ObCJθj(OpO#أOLm yH`r"e$M!Gq$7<7W'])+k]DjWL8ŁT ;k0,Z͈fzF<=|D([2BeHΒ^VCq8WtLaș&h-yٻjdP0 !VvZE8 RL4>Z$SԢ,$JAADP$ )o>'2Ygxڲ#~KQV#b,X[es#PV,rr&8׻(FCE6fIK%g6(QAQj/S+CJ>["26rm/Z8Ѫ>6xg#,$z 5tSLS)qJQ1ȵi74;@6Oђ`KŶ䁧}'C.fF3uA(hgwr Ei~g ZKCh;gAtԜ\%Zҵj΂2E$!)%OBi76x^90Gl)E7M] 8ҒJ`$*hW Vl2|AmY  ) .y-VlZ8E@ ֏*=&٘EFlɠ=p]t WَBw(0z3VpCٔPJKmQ:+XlPXc7j7ɔ%;Xj(![~ޣу7v X:B#~<䰊)Jepe>Gb^\Nfρ#eV:W WǺ)kR˩F ֝ 1wtufU_P?Pqe+5-#Ͷ(K %U[IHA'ua˭)e,LA$FT7*Gb_N`_`#yYSFG.3R5 ʴD(̌`DdP6pDV|QP!`}qUZ{-ja]{" YfM,++E- 60sjlRȩQ,x86ᑔ$#Ljf ܷa4hwS4t51Sobp(Uc^+8;= AO(%'59DkӣhE:;.*@ڎtfCVUVHL b8EJsCmH A+RT֐5CŢ>q ͤJ$`>&U0G袩S\KSz7E$/Z" &_#NG MLPnfN̓U^VPm5J2'ojq]nc=e*l0wݲ*Z3%n1 {l8e"$#ؐM/[%m,42QhS If1"!I#ۈkF&"*8?o|*Z2Z"cRV^0fոu`k#bF<}E)$AtY1"j*"('k$AM-ɼ΃+[V@!V4Rdڝ/B ׎aU@A 32&ֳD$JUi3dT {dDB32ˣL.:ksV~$YHO QhDCjH45)X U/Ƀ%Wzl$ `"S:th 2_EUBX0(AN5W$SZ/Y E[//כ0t DGHtⓎ QY$}lr?iBG9P=2ɶ#sW |Dw`cנaoVnO4Y狳!찾i\}Mzj.{kx~?Vo.6\h{˂C?x0~(֟ (y<C9e~~~~~~~~~~~~~~~~~~~~~~~~/^A0XB9N&o1 c~7{jC֓bXo٩o˴֢9E Rx `/ 6: @&Du `eN$K.gهe0i:JZ'seWrl{s:)EwVwy1oX{NTc~ rYxea4yDxf*f8PZ<,{HC MrzĬg߼ hؿnaNtț˳[Ju*Nj_Oi1ޏ|o[8A=mk>€6n#'^.v:Ɍ\ZNceju}zQjymF47[?> ﴹu.~j7wv6/,i یh3ǐ:3`X7K\wlJ%a*D}j:=">,;{r`ñ9K>bX1<3 L(]]8xӋW յÛ=^ȭ^}M:KZM~~MIRώRJY kmH0z ,p޽.r0pEXs`Hً XRN=T> ;r\?+XR N$'On2zŁ'Ez2ZhBd;32^[ة1кOad`#+/>̦Jpq00Ӂz*#XRd|lOx~$LG`JھT}OyVA V*[QN,:`X;!zX䵂}XVk!){"dIu} XOF=g3V,6zrajv7'oN.'۝jޜru%i 8(}䙞:L_ڃ_^L!gʧDp`ܑW@JO% 4yqS (񉀵]a*`2E!\> VNIV<xM,+叻`Gf"`-Oe鉀mXtPWweqn6zO &V)w줰?XҲ=Ll0 X&+f"`]66,DVݘSBD*9hz*-+ *Nl*̰#S`ҲN{܇{eP7Z32 lvrSg ~pe]zE")s`aw@+/~^V)alo[%0FR񪮇tuY>oq~oZG=|.?}Gݿ;Yݦ['w߿-,Hg8-Ի|8Oth_A]W-8UJp}XqW'z*bϗ A[\.W 쑵{ByT1OMu]xyq=\DyRjW\nnEtVWK{TQ;29ļX8;qf z7_{c>һ`ݝnw6jg)w;mfxS}4>%dՇXS`\HP~a>hq:aSv<i53yqX\Bs(E]zawe*AViɊ:*Z%)ȁ2'e(1%zPZBiˁw:VWB% i3+# /hL"MRChʐ/~m>m-=k@X-oRU)6%pKH|DFֶbג6sb4WCg(pяYOy]|qB稵JUC2jJ?G`%pg4jzobzm4Bi|pFtEP:P֤FյY>445$@P1BĊ@\BGM[ GDfժTS*%JJ 蝧 baLbS hMQALILN4 EzR^Τ^oWe Z3\Q*! ᩁG(AX=$QQ٫՟u6c)H1h- b)&eGbX|~'Zs=uus T&u@1\CIJln8'.^lMgٟ!F?:F"DY,Z# ޏ$B-k^V'`з2i[LamakQ͹n^%t/ "d1䴅sCw舌%͇u p1 HE܀]R! S L4 Ij, -pJkKuR )DB[6od|v`DWvԖM!ja6ҟJ%ioD 5ݐ(S y_BnL2w) Fm0&XıknctXp`ldSI˪qΡU[QmwVҾ0mѷZ{$ݐL`01V9)CF*ₔ ]Bcؐ~ .Ec}jE| f! S߹_C k` )ʏ* ͉-U2%DG >M#:|&aUBb X8HUxv8JR`:l6j R5ˌtFoS TоrCsT"ۣ,J]XB| y!;(Z@O{1Q+}キѵemb gյ%0@xG|jWbi0xUUZQhUf%Y)GXq;a[`@!tj"ظ&( Je (oClNf(E[G;HcN/ͷ}2f0>C;zLb$ e$0pHy+7>* 74F4ڨ4"! 2;IURA>7)`A ĵng K' O R0zfqprD 1O-eP@rw%ڨtqw@eXӹ4(rVhRSUVA!JJ!?*XLPF[NeEݧ٢V&ϣ8 ~Tu*a%5j(G pd}nY)̜BZ!A# r0X琄aPMkV뛞j R P)%Q4!}z Eg=0H RW 7p \삁pQ @",Xhj`\E>f B;mZ/\8G jyJYʶ` #¨ >8F)Bjb+2d)CiO(۞eX<TFagFeg x+KMжt Lp3p3I/8Kզ%jͨa`j){D́AW!LBua 2Ch.`(rm [)Ycj dj/ER8;4CeH_8|̹( V@5!"+hI 0]V(.ZWbӭR ոkji^ `Eu7*YAYDGbEg#2h]GrHF#}Od.mhj2*deH(Nꂑ#&d(W#BcA~UT[ *ߝ0X8Jqm趠 gߗ*ͺ piV%)Oݯ QU+qi&:YaĸH.=5.fyCuפDF䮳@O=ByݤP5WdhZVC“ L>R!K a +Kw0 #X>㆝: |֫~{ݖqrڬ\/C¤6='a<@B,zXsnU~iy\::GjsuMڳs'އd 2)me_[M寳%n/E׫!Vz˻b;?;CJ>%c_W|zSY*>=ߘ_A5b.o./RJ_rXPgC_,WlkeXA-a4J0X"D"p~*`V֨&YX/A4맃1;;67Ap@g `9of{ k1;lv[B`V`HL,9aT*DجvDRM, 'DX紙xgāfO0`y"5H͜{`V gQ2qkxC~}yy5e< ߲,/|uW ney{t7/Vb/OONhf =bޕbyC]M}<9Ջ\T,9f;7x}fkNƐZ\g[`h}7[/r+1xՒǝǽ}A˓1v@?TMxV/Q,^i=sQy!ULQSM؜TDn"nBd-3?EX($dYz0,jCE_,8ٻ?b //[i y>~#???ubKbW4zOGps?/TNj&B A؉^ n*-y9Tv9l 6ŽJɇ%oT0xT@{kTxG&z/<=XկϾ$7\O+?WW7yJ?ɓw?|GVaqu)'7-) 4݊緱GV~ɚ* : o,w"CJr"L,I\XuSiYXu*EMU3;<㯁.NN 8zu(WPp?߳}ӷw*.m:^uèz[;z#,;+wP{x}3vFS =5;]|&R,oŮWpsxlNtVkχ[0_+xõUo#W-yG|k~ao%}qũ! >pV;8nfLAx4o>v_w+\\Q#Cx|,#uo}wl7Ϳf|3o>s}Z]> #%hAAd:+_eֹjT\S_gJ.u6vӖ)Ѽ\0Z"rڴx-rQhL֛$iꨴZ"_)z)<{<(MoSn /DׄJ'9T١3EQy6O:Z< A JOJvZK)'ҘRpՔs}GX$ȟ\rMT1\&u&BʧPZpFqADk AEKV*]M#_\R(1PD4P+rAb-#,jYRU;)AHE<+FzdYĥB60ZSd!9S Ti6d0 5LOEZUzz9oZFn٠5 \$EJj,$Ċ~*D٪՟u6\xpelGG?F*lKEXIZG@V/D؃qθgx.wD zĤsNG9XJ l8\6iMf{O4 4B@:FRRZY\HD!4̠(XK-Thuom CiB5dlܛ\3t/ U"sΫVE興1Jb]q5$D"KRuD{-H+@Ђ0䝱I'5 0izqp a^ڽlT Q7E1p}@J%bcb[ij M {-Ɗb5xkn蒱>c+/l*5d39j+r4wV{pѽ0mѷZ9V LF sSU)T` \"Ɠ/0 TW!ho%PE@ KL0@(" [mQ`=GPi _|Ʌ3h*C܈.z^ L1Cވ$qbofy [ECF8dдʖcO2t#13{jmB)! n7pf@$%֫M!Xv)njM){D́ZC@*a-d2G]]Pl8X)ng 5S`*g0ddg 3@os.* ~fkEDWВ" d[HfH5fʎ[)%\h23DJS]߈l.K!ZdErca@lCY]D!^z8 82$fW_cHɐI<µǃ\BJ [[+DSdR9;^ " gM`̑+ X%9+Z"n-)}p` 2 ;U ~; tk)F S`䨌YM`X@q`0zB(z|ӠPy1ҭ&,'"WTR ea!6o~`1"V/SzPH)+FrIbTbtJEW4)R ah N`\)ɕŕ|FBMDVpAU@08X<+ Ĩ5xe1TDHV)lªj?ẅ́X0[L8Ѡ=F8Xp-kUdC2B !%J"2Ay@KkJ^%8 L`ڂp:xfO& SP\ iE2nV:J 'U#M@\@ۀ "%oKm %.VTO^\G=ie˧q AWa3ΎF1gWEq3Y:%dx2|;LH|";>fE9']M&w凋 P)%3qv1x|$i@$,!}>}x(} K]NêYo:,./HH-;yNR RDPA"uHD R:A"uHD R:A"uHD R:A"uHD R:A"uHD R:A"uHD Rt:q'lH:T7=u"D׆L9zW"8pב+q nĿxfjA9~!]j/([mQ,/|dU m>?~uqK11˙ҹcV0YF3p*3fsϝI 4m3 1KbN5`vZ%\mS~[9K2tzj\Wiy+$~xRmO{[^oQof?&EϪJ$dv<{r8_0UJg)7*Nw|r,N߫HKntYF1<\o"x˷ʫi % q^q8?{m [=Xբ 9c0#w#X⸑WKQ#wGe /u]>L~{Nj0Q&Q%N-}x' 6MndyK: >er}ϵ^z+X>Z~KU<sa}Ľ]on]-(«LFWruft'0AټBIϾ{e!3 $|)nAnGK5rz? QJ"s _~kK]0*9֧yZ(s;\{0砠JA1+qZ0$upJ2-"_ErL<~Y~i񇝠 ]nQsf ˦Nͫm~d!P-mE ܚ&ji bVW˽7gJa2*m>Mo*L*q_..Dä Ks5DeZ)M2etujJ4Kco)3oenGFHcJ#vyPAɉm2d=F,Xo"\E!vR,w!ʤ[Lc2sW=6h(!.؂ *|^>k7)Lhv̕6-v)-=B{uݺ.ܺ=mIkA&Fe.IfH2E,IZ-_p|7bO57p",e:GTL˿M+M".K;[?z(@0tꁜ(eO4$֗M!.ueӋf\j\,LM?v4û;xۀ/CYhջ}ݧ+sYqƕ$.=_OmxxJ'c ʗaK=3kqQ4t%x$/65XZq}CX0U}7 _r7Tn^<\hAh!_xTcЃ{`bvO->{|x+`LtQ|[Nt{ɽhCHeWh RȦB4J8٣G  -X"j0RF޶J6Po83\FcYПlW-E[e}ԙ)Rh)7ZRf3AߓG9:vñ{uI\; ۅΤUG%)Ɖ9c6A عN+έڵt/v %v|=WW^>)nGtRh|9j(ߋOFY۟DڅʰmX[Tj0ҊH({ߵO.~|WՅ{F7, ^ӗQ2u60R"-luʭYLhﴴwf u&><27O#ttx{i=}:(YE +}gC2x+:_4=E_4"T.uۻ ޿WXN;ƾǡU ) _Uߵ|gpˣG?Zj; 6bzuez; ,qFS[ĊFshD2] /L=P%pvARd0č=n3Ap1/<|7zz蟮\ƭB=&3Q6NY9KH0ڸS fL =U:ԂTtaD|Q\<R,WA!C&/1}sIe=a昌fbS˜D.V2LE1JZ rkb+i/tbQ*x T-y& LHq$Iq%N0F XU/d^,LD<B[4f氍MŽQm."  l%LCI .`$6Fr%%ƣ+BOʹzw~+>:T:hI1)" ytrgV c̹gHx+Ipy6YvE1Ɇ* Dyk:z!?] 26V0lZ#\ dNA8& ]Dj!cmAA˶1("fU]ߢUΎ>Ƽ  ]_N/+4r&" I`,:y¢LNyE:؁] vyO%'o=_]^UV9ț *Db:h /ц/R\'!:^YKOl ;lKZWbG1L"1R D_@\Qa|f-T{"ov{G@\ju0fp{ג.`*iBc)wp rjn4Zktݒ-t+ƳOV`8a:Y"<fb[%ZUGeV1zh#ھ%\\c? pO0pDYp)C&f9q{9'f1 sة.;olͼg)ш 6* -sH%ڮZy_POz_ǘ`HD`7Y! ZsӮ(v5BnN$Tc9 V 9J*&d6)G2hE >4=kiZG8LnmZ0KH|x3Ip;& )@~nQ>^? |Ӡ.TIaJ%I#QTY}G~"򾐛|j^ǻ/0pwJvG2 m1+^$ H$'yGH},b%UGRҴoEWa "bxĝwUԉxaف<Ey_P5;ٜuuD⽡!9"2'c)Ducǵ< Owzz KΥu1cj sa<W@WE&)b(rxBlfdH*% |ȻqjY1pѮafb ]Ț0xP*`$(P$M4]ЦMtkW`@#1X{IVDOEmRJ+xr[U0+$mʳ8OF"DbaFy0tmQвQD[3u$ZNGʻ D#ǠB} /т'RBTl:fMV'L ሣ1 "Icj+ImeV/frZM}-1o3i"^/Ƹʦ99Sg16aGcK$!lOSKrըe0`U!9Eac 9#eUu1]+ e_n2fbGySկRE';xANE"r._3ƻ =w`hIvy_hHvdx!t|KE:W >ffWZ+lIcdnE:bFaQ͵) \а]'Z.}j~X|&26Hxح"(uⶏh"򾐛r}ݕ8>JGudf83)҈#nt1 28r:XǗE1=]~~^Qi~G㭨sx6ۡlwj , "EX!+!,FP@ tG'!)6$E=q'd8&y0O C` B j ;2nJڷ{BM,1.یN^Ng،z  >s9֐W0xv:SIk}+N>i>F|8lԺ,Pq 0\9 n[qwpOzmG@2b$Ym8{a*N!Nؤ5*i." 7\*ð\,`!shYu۹(" a',UD޻>;8ߤof!Lz;7.[|OVkxo]b=+lo9*.7tF^/o.Zۡ]Fct<ɏt!ד ^;[eq |h*Y [mWD%}wIyǖoH $8)[)HK.6_6|~)ۓoNA[Vwx{=v6O$Ot:h_C$&ջz8mu_7]Qt^6j1oˢ{;[uҴ{ s `Vy~} ɟjFm l``P륏mq?-Q],ӽj:{^)Z:;A!*/3kܒD,JSӋ|Xye# _BL#6YTWrN1F+  a0UIz"%SoHT\>d-˂F*sR=N6 ~tK)UCMc$xɼM98X1?[X|}|;nh}y9zr܎[o9oZᙑgHNw?ܑ|C 2lڪeE޳FWK*=X6m,dswJq=m{fjwrj4vk"(ER|w1Y]^/ \6yumV_in.r{^+93ZC Nͽ췔8z.]lϛ?x]{805MumT[UW>2aqg`]zdoE~s @@&sl2OW }ӻ?S???{wןw&qM{l\?͛ _n?8 |Dپ".e4uilnO}[Y,Ks<˽& Yk\՟oNsNk|Ti0NX\s!J@nL)6N:*ury\:x0[復CJ:*o~­OרޫIQT"cz?dm]7u 2UN?vmNu#eAo:+ PW,T:?diŀ?<ۋ?@gnj1߅Vc}5__l^ 5ؒI9q?df,{o/ ' a|H&}f^~~UF&k8}=!eL _2#c`MkJc}i6Ci6-+|572XcXag9Qry]cvO79D)jj*L:Ը$P(I, ܥ(@rJ+q`:UVs*Gqs̒fO0x_nj>xsRYT#Vj&G#  JKi_h}JAE%bdC:0K!R|ŞtYߴ P}Y;\YHpdC3Q,jڂ>*k8b=W =18,ǁe9gcDhf6[;o&RAK 7)w8yv:TLLp̺&xߏe3 8<Z|IW6k]uul`cV%4l Xt Q5ƞ {X.j}g9~F˪I+x:1OXupnIU5a įe9fm*AɅ/V|K)xp)gR0 ke̺AڪO[<= Y8rJZ%WgFhXd9 '˛E>e:vHDĘgl~(ɇثxM%hQL5v,0BTSU.Z '2fQ3ޓhvQCJlN5UogT+*n rW>\ +0d送s: 'Н-E&N.MA8aU6//$6XLGd3ʩ)E]nX@3ќ땺x F(8xf"^{56$eS? -7)joBͽ&\]WāDx(HO5bQSQ{2$EmٰmK/3P@B#&FA@,Tjr՘ V{Z0V jl r6V3=g뀷PnQ4ˍ.CEwƏrGbDkE A2"I -M0kDEބ`u?uXߌ֣{ZQ^JC1Xm6QFHVJh%N"5MC?kQ:$D%p5{NQ])/ pB(QG 5x>0b]FdEPj+/CjkJ.C]\ɔIhS [!V6P=ykxč~;K_[kiێZ85%>[P T'1L#6eދ}"ob K `,gE㜠N&sF4XxO3>d |1N<"c c^UQ),L!0pXxjI%j a6(lϟ렟1[h<82>0}Y.՘WyқcRލXUb9^QѴ|^ ;Jсd3Ap-x,qAE%BԌh|úhj#Q h!ь{\ALx'A W]5cbk*be)w4qTr1%xQ|rT#i|\J5!(/%aBT#`y%Fiҡ#* tP&V(-4m,aèۣa;SBu(ڈVu߭sk#*U! )VLRlylyxOM_:F8x5 _bxHc0& "5W1`ulصcj)éXdzD|L}?ˀ^TLrj$A*4 NCPyQSRu)xS`gCf&)q~ Zt)&`(4di\8ĩ FOLm(*JCr%a>`4kjjԐ+؟R#UJ`ZhX+9˭CFF%2EMC=o#7Wqh->whTNXp:u%q3iS礪 .ԑRG_gA!w"06c({1`nήvEcᒉƪv01C .orkԡ8*FY}\Qw_r5`-z1un>vӹ$NkA@@u &%A*2#O9p}x_8pg ̄&_\w9:/iT`7ϵEK,+.'˻m.,q )m2qjoI,A \s2Ծֆs|oTǦ#nY0[]M:ۙa/Oۑ_}G32[ ' dݭ'kr{xeއJ2KZ'T2O1)Z٨chZ7튨+(<γwDo3N9gw猥ه&%kܺ("V`!*IoBH`xfu?_^oBeܕLIk-Rd6xSvvv.UEr:V(}2S׆j~B LԂ$#XJup^FB㰩YxGVOi !pݷD#vFxlM֯m|=-c([] a ϻ8>]/(#H^H|$Z `Iύ2QriKHްel=w<_>Aߦ M|u|L;1+N@.!l[ sK``}qhHU]C>Dsz}9 ]AeOYvF ia>SѹDwHkk9RCy/[ߟ"ء\Cg3 0350eijį^#^/-tRot5Q^D=^qO^kL ;^`&"74iQ][ocGr+#/v/ Aj8Hff'O"*<}D-`QO}]U:QDB?Sӊҧ4j>Џ8Ҿ}oY#;2<倩2sJE%o%)OmжAoyoI@) Jԝڭ;İ:)6ejӦE9&/8,D ;9KX(.YtOmQ viy>&<=" &>ewGߢmfs^AHiO KsJi‡q]igD}$wLpI.RQTf ŕoŧ!ҋAE?o!YX=w@mɓC]qj&ͧ}: %+cA_ӏ"}#}K't{Hg ( )p#+9 }ͳt }]Մ+wWTSgNhS@b"9!t;Fn1.O۶Z0cmPBwZS`z ]mP~pq>NSamCww|H ZN*J< 1u&EÚb]WvMZ"l% ǃ a0PScHL#H!c&.,^ʣgv -ڸkO)'9&敒$K&!>mĉ6j*X,|슺o'똬1cxLIeB%\4V۸(s,_q oϑշֱTjO^ ఞcbDx*4عH 1<!s<l.sE8'q1udLEBRH. AirY!ѷAiO/`(uk$;4՝]Ҵb4Zuq vI]41EZE##)9J&y$ d6dbq^< 0O8ntE^xSƕA'¼4Ox(Rl\qU]mYQQ,}km{oYWT e*Uz\iqx%TRUqK6 4A(%<$iΈèSrvRVk bNjh j JmEH< Vĭ7+1co 0ofZ7+s))D}%NL$-1F V^ög7+7+1X[Mcf 6ȉ(= &z>hDWBy!jC X.sg0ջ<`fYEG p| v+-OcT;F- R# J U;Ĉ[A7o){gEҘ}!b0Z(U{aP=0H jy|:8Wb U*'PAL)8wQ[`+s1ҨD˪H/LfB(:Gpۆ. ΢:BuNxB~A3Z+@u:iY~X5-ʩPqϻY ™1oxtE+s} hp-)4sh^4Qk>[upp.+vEKsn_`1J,|9_zWOch];+v'dі߮\ M"ރA ! ޿36ς!~`~E բK^[zIjcvUy̞&{/_N.71"N"/As态FDzz4I-K c:`o̧>/ Y,):V" 8Ny%kV tUL'x&WK27nXKŢYu+ھUNjE,/]<0q*W[V=KEkZ)A0DLF&rg@EcB*+iA&dF-+-z,Jkqm[.:Gi'1R~t)ұMx Wa]_mR&^^{/ٴ}d3v[-Wۏ7U}d>n1VWZs˛n9gZ9؎' $T:EyC`-D<)Q!ҺTz"T]\Uw7@>i8պ^TA$f^I;B1-, CI"UEmXmq3?4>Jv(vQ5d"l̦~/}XoUO)Bu5wcV/cxq@aNjd6|?Ys^nmY~݄OrׁqKU.fVeɼ' c5׋ oъ }v$ӥإ9ݠ̳8FZ%M*)*& ;zU܅(`B AgAynC>϶BB mb́ fɨ!fD rXIH9N,j]{ "d cTE` #crnF\/{-q0xqScЂmc>Z/H")'i%-W+5(0m7$ד e-c٤B*(ݮW:emlt1؃o |tۓWs8PL1ļ:$JN"K"ScJQ,c >ayG|°r/@k%:$>r!T)$S.yc|L<0J;L.;sa21sHFmu-wL| 􆆐zrăhr .eG)jlkK<[4$q]$/z:F UO7{yyfqW_ސ)wjcHMOX@HIj6Ͽ&*̚mAB\-tQ54Q7o..7MPRD M'jjS7'NCrS4T eRYA,&)Q"'6Avvy//y;xwqsG(3]J%%%)ŽQeRik{P[RBgop}- c[?RSV1u{)HGkn 6G=|uzCD|=APDI)GQP17_#-􆢐zmrej\/_m7S h(կFKq6bHMc+ƻvk4;p*B> 0$Nb<ĸwiyR_;f ݩ5 =-cHLnqoqN0cP0SPAZ&4 2A;j&zA toEQ(_nU h=R8&h,&'+{s0*!x`4WqVJ1g`d.<0 !V%59{I2l'}'xJF3Z^.Kc򸶓X iF~i<Ӷ f?m0,5Ӷ<;KҘhS qm,OE oVK;Z5!5 70Z !$8 #Es֭]~aDWy8|f'h)lY zA4޻f8U1ʑզ`-k,_$-t7`!PhAQJ(I#s*IOdɾ蔫}@hXgif,M1?]SFJIPKMMc!M9N\f4L[uW[CfikV¢ |w\ @XL,ቩ:=tZ<2QJ[HyU H-\w) c?"0ʩ察&\tN/rQiJbv/kC-tԈh 1BuAEbfHӜzxHxsy;НoX [Q(JMH_7N݂)Bw?0-P?^98sB"j͍f\kY|ճ4bowϳ7\Tݲoru?7S6 a;QG}x$ ^&Jg3u#.yp@="]lj$]V&k'/A&Fί{$d1yΖWZ2a٬^~^4mCP(nb' ۜYZxR\ߚً? {ͣ2{tO\}?Xm|N/uY._uTFӛ_)p| c!^miɶJ%Ж:^0Hj^I߇]6Ivm"hwmOTrhmiQW/w%/| Ng9(Xn33Ŵe-t0TnTw˅QcIOV,pd6 t0Ivl,k#!$+BӂW(GG )4Xגc1$h6^~l]@w[x08Ac f++@wO(D^h!}%SWh<@VSTNN"(fX0/l_i.v2ĮK$bMC Ʈ|a/;AaLDFu?PL^O,ZC2)I rw3? fY!2`MP9Go*WBnĈL4bMʟ\2VY"D?3׶\3tv^>xB?!{:t ˥&ͅ2wo̭ks0qJ#,>Q9 9%/ЁQ#M4H916}d94R<5xZaٯ[=o$ו_N>e[5Wtc4&j}i2ݪy2AW)qf}K7R2hT< h/jQ \PT1&-L}HZyFc}HH 9 N}0TiA4ؼ~ɤ=4hX~B<⽃N!S-# LHqQ'<!*AWpWg N|CkPQPbX^Y,v,\S 'L} "1Zɽg)Ƙx 2P?|^ds}VM cE<.#;Z++ǐ5#r1.{!`Ģuj$NLl͈{J%[ȸ_%D'V*+԰5*?90p+Y:IҞ}H֣kdW{S!ƆO kBwwv;;0kܢ18H^1FE#Eܤrbu&P s A`sP8a<跻 oUBa48Lsf-\A+av;`=\;B<]8t=Ycml먇E@V~d=R…)?6&s9OJs JI Y.D!zCE|=AB AEK$׵>Kf[ 5 t$DO݃)> >}`.syc0Z% Rre))/ #s4,/F .^;|ȸ|RlGIY2UI#Gr"؊dɡO7a߭W(&5.N]˄OuM M!Lj>#7d/JOJfwFn._gk;Igվ};K#K/ 5u& zͩko} @rEڌr,  ozz9A뫫@LP؄Utv`Ib^䞌 {m$1/j ;dnf/{ݞq?zflv[펐lƢTXU#{UIfL)2YArnM^8dM{b\~Ž&\*̈́;4 V$0k zW M*&Y散T(0q+Ebj_{]]_UByoʟmeoBI>_wsWWxz|?)gmr`x]y Z=>N܃]=lXrORVuB%%Bo*l+B?z7sQEA)Of3 m9q_;EAT0mbM bRe06ODJ4(*'6Md FV ÒPYJFeIWIHV^8a\Ip 7۬er^*__eY3AԜMpLxeFi}y,83P 3$%h/Cp#QӅ;~ͫng rӍ)z|FM{,{L{pQuSݥn TDhym襢+GB>[G~ht.n~0Гl' gƬsgo0A3'Qq;^~} A\(я~|\VDO|JFoj-Nq=6OY#DΝaOhUEOw+|_~؉>(3*T߲i)lXO* QQgGfxyf<}g>%@~v;YONjH^LAp΋itTwit~s^GgϤ4ycT5#H\#p}iqK|F$XfRZ'@VfDKp73T!@U19:aniw`)0*7!0>~%qN$*@3yH7'iGne /]|zcۀ-aՂq#ؤܝ\ȴGOZ-f2p8ign (+ 2gEar֡;+wxw[+A3Hetn³3h汅@-߼+1 .<…)̈wݎ~(~6 RT!n5. K&}eJ%5tuw88j^jދ7~?$?^g=.qW e2#9('%h@ӲfxLps`SP#ճ!)*CI$q:@Iq޽Xm MWޘXNzBhvhuB@k = DDR |btIŘToW,|;R=RbUhcsZ^sz?bSՅVuxw T?Ͻ%gO\ ]4Wk-Mt*("1{^K"':q -wӢ(tIsUƈY6c7Z1+8 qҤ^T,tsgI˂է5R)a^uӧů/h~1߆18>˟Z՟AȂQDsyikP{/%ga/l.|*֠U>V+a ȑGIj&ΰqESc(EE*'DrӮW%xVZ2a7xR}y^,NM( BgɔPBB蛸=R*ztDE!w|LM#k]ä;%837902@G P )OK%pim9gŽ\'?C!]yGqU^آj1k-TunfoQW9(;vXdaJYŕEv!Wnc׃FjGv^rS8OSX+@Ui|נO}x6ڵ kE2͍=^KWӄ",PSOʼnBYQqfj.є^ޔ%ܗ'iKwM ̍jn:Z¸p|֋uy hD6~ِ۞}'~,ŧMD̶ >Rޯ@Ĵ[WUXa[bExDŽϟ ňh gph!8ʤfȹL8Db j"qT7ub&N IU9Б#{|y3eZactg 7DmK WE N\Lw6鶃vN+YQrXڤ@ZQSPLz2K ;$96I81} ; 8C5"d/ĉҝT:*I[O{/dަh:x/hwTf1 ʴװ-p'3x9Rq~MTٸ>or f/~т3HM4 +Ib؃`׆AYsjX Ӛ,d*DڨdLH tN6R0!5K$D°zO+P2+4kXkk߲[Gvf>B[>꾝UƺlwXkӨJn'&YWQeAYܮC<5; 5!dUNiDBc)gI14R6u6ݵtyg>qIP (Bq7j|Wv<z-Qy*8M3LM |3ed&^IqVI<;1Fes*:y3LH+~s8exf?4xO XX FyESLm-3pc1w,Č50ELq /Mܡտm !u.:^dG~68t6]=/6|T3t4a@P9ۇ;s`01o 1-m9p7Eۂw(8`Vle&"7]`Φd=<̦8#qNث!dTJ&H7,ZZ/g{D%]rQܫsb`%ձrs@x x_yp^D">%e狩@DM&8mك..D;x̒.p|.vpx Q3'͂w!8fJ8+,N0?{ᅘ‘\Sbf{iWFኅ:"ƺG؀fcZ>#r"02=ݔuwyxї&n ;Tc'n"r Y4t.\6Խp;ߑmWpyaULG!=lk!Ǥ+2UWYX%b;߾A%S"@GBUWQneR Ϡ_S VnxtSL% -E46f-/6wC$3kB2EfH1jYl.]iMyKWT쭲&O\VeMu$XQ',_:YsYƒz2y،jn\gWdCl/ gU31ܨdI f^g )\z/ZeY*OHP 3ép]B޹ަ D;ya%JDSJ"(B=O|$5 ϼgJzI_t}/ xC*(JMT2I(Y$#ݨFx/#^{kǣ֘66,<2+]|y7p^'b'ӇZ@z_2P?WG /u0]3 g~zio?2"v/ouFكeVoys|fvQjr7-gwLKzA:1IϤzRVѓfGRMlRxVI`#C4R:aiq<`"'2YbVA1c*T%g:tdLN9HTT֓'yӠ%Ւ%ВC5}X_@NIB| |HL`C0 J,N ء/kB=m8`cM>[c!„sۊphܩFS] ͿM}eMSHX ݪH%\_7$J' Za+ba\DN>yz֭Hh8k[k:c u>Z`> k >0OTAw-^V=a%v7)(][C i;us|#=eXT'F|-'*{d}e{ɫ=)ؾn%k)),9z:i]$1xr;]4pOZ dV9$_3 'FX_;;!LZbg3{M **pLِ9$T P昣Wnz?sv"sků1mH<urQHVyXn%JIޠD Xc1DXk2oQ÷!%619UHj~iQ&DK6E3z:/3VY AH0 vUR pJAx#18j%k4$$,ѰQՊa0}u0ĀcNX\k$.٨0I/kp{!)4E~>SsN&Y#L#.@V$Ycb6ǗǟM"c;oRuϓU˷F 7/ pO0KDYKGF9"nFHuӭyT2%FP5&BhDE gH9Cm7vN7S s/J 9J*&d6mmdx[&o{jfIރ䥸 @d `@A=Q`hLDŊ҇{{enMS0*DjSR@ZJv@I4rT(qSn+# #-;IqCoo }A燫iUvlۥcS%g`6E>(IpIh!+8o:C6Eғc9`I#QTYaQ-F)!"Q$GXsJ|$TQ3 ݩSLzSD88r`ac`ɩ4|f:4+{+АcTX!E}J*QC=2Xs!ƿ0&f\\`]LvR,GIzBzeHv4$EEs@OXb0w#DZ0;.S4c'@7 c\QYg/ܷ@ʭ@ |^pHhpR}TEJBn^r'*njǝ*n:/8'[V/r.aWLWO "o/ }!Igkc}G=ay|:7\[1Te~ 9Xq߭U/†LmϿzy˥^;. >?~ycSFDZSA/F g\58 gƸ m3r[ڑ3@9NMuMͰm~$L{q?;˭qy1.]q!IYrxp=zC׿];e+f rsĂWk'uZ%.+Ep\ H_X1Zqv{eד`?q@h-1s__hOJv&"| ?]5zZMl];Q>SuY ձz1nl[q%,Ysn*6i;(C STO?Y`Mjx-M\3hF41GovΉz/%EP*>^FR1p y-WOn%|63.Vѻk0kҤxac؍ ZĬ'l˦[_#!c6a$Qد"m q6Vo AneDnl >F]Qkx]v O74wƕ`h i6NPg4!O@wp,1rdZVA뫶(E:fnP Z_1 JaX) I!DG83!1J<[/8m Ǣ})kBg/6G/p57 7F ޡyFՆL, 'C ,v du&ءy.P0VC J\,*sh T͗=2c&|Mko85wqFi&JF[pwwXX0 7,Q W,fv%18G))Ŕ3^.J$DpD N1őn`Q~&[z upLl:W1_MK1$@fb '-G6#m >^I!$KFȻ@2 2-)+k'2N 2WˈE)\@Ab*DLgġE#@?EMV:#b-SFmBl|H-rx0\DvV2#Xyhg"G0-!k'ސzn>4:8P*@pLxe0͎;v9D,ƈ#Ja#Tc#)@|}lx i S,f;V4dcƀw>0#=MKF:Y 1XT،d4Bnbw?g4G d}A&m[oƴqeVOr% =2Mx~FQӠyٺG%Ui6'ܣ(4M&j*zd4Un?95@=2QjU]6b+[0\=a#ؑvј*-zr]!_'HT3$s$ h4S" +u`'ջ_܀+u0 V< @&K"[*V2ڂNrzx:i=zఴ-7:3lcre$ljrl%Ҙy64Qרڣu'p<(Ȟ58-j%jepVQmX<ԃ H}mL.;F.󘌮WQ\:`y I(L l8HђE> %6ʨJI8 E2rmPT865v(.6o :(usL~~]`#-8h7UC0˄$H)!HXJƉ*n*aR=$F*5yM* fV&"r="\QjAt8E[ZWAXD07Ѣ\d*ѺBxf!"Y8&Wc)O Tԍ,5 ST g^= , \σⰒDE*mBn iO>Zp@w`crlhGF[p 6vu&z8,Q~cH*Hu\2:@-5 GL: D1e@Zzd4M0xvG -}˽(4nLsTNI-Dj{ڍ.)78 QO _ZK/GFcpt}y'%Dp=2ڂCpӯ!8Fg-V3AGD6+!x ]c5t4WzwKQGFcpjPn4.. h E01CJ}h o͝J7=2ڂCuUE(Y&V:0Sg<|9MGh4W8,?[eDžb G7Vc,l>)`it;tqGgnƲw|"e>V)tK&%A1**XJh1O.@ A]󼘚~5Wa [$Yĭ,Ta.kq(|&q-x'ZD =D$0a"}?+J⻙K!KlO],U|X5Goo)7`j,le lM) {pWl緲;4Sy6Si^y;ސC@[ʿ?f],k@1B|fJ΃_v*[k_k~`/+'ЮK֚SySM*G5ߗ6@fgPW,z \ &X+f 4#?E" S(Dt؟v*Q L nOY߮~;Wn 4UEe#3~ Tw4`N(r6ٹo/m%0bl$w9 (瑅nU 4~\}뫾)JiNƿT55YWzqY/ӊmW8 jZ`_۾uUIj'5et暟_-/UkZ`[<؋ۢt^*?J0EbY>?6_NsC@!^_壼o0&[JuhVrKp:T DXްzdC~\p6Mܼ(ח޶kdy^N=9߱t&P>tjXOv&ð۠; 15vQwa"juХ-qe'p_t@i}Ezdi0i@Ň7,]-!uw*=#K"ղw@SYkd0 |t̃$HfTb=q̦]O${UjdQx Q4]^z#o,Iqv%nヌ KRs<rKwd9`bZΊ)0mVvY*,wW'G\p[d3i? Qͣ^G0j7|+!}Md|V0Mƀ/aYf }HiȢbg(V0 Cqtĕlx4d7 1j5j<θXC*clU\][m1p7]1pP9ZexL{̸JxC`TǤQ8FB|2@QjC5D(6F|丝fVYCeme^U:tݺXrvY8<}j㝚2S3mzk횞~r`wȁU6FhUΞG  Oopn!W} HeǸ¡QYI֯*$q0Xpiq[3H Ǵ"/];]R?\0xH-2Gj&6 WP-u5VO{x,>zv)k`P ͑?#ZC^6+Op:O'6B/1CyC[%ذMv3ҔtYtDsjBbP!Q"@f 8|WM(˸12,˚W wO~ +hӞv<}ƥ&~1ʈ1А$HX{~+uշSL kvSV;z9+]ltd,SqC_#pC޿爄R3*m`ؑZ{xܷұӂA02ēTh #8#~?'{c"XapR|ȋW_*tGDi=O_e4z_ͿW32+'>Ǘ-ֶWF"jP3%ր%Ýa^Y‚W(('_ DA <ƿ mUP6 ]e+xbڀw oO5ou~QO՛׳4U kNҼH87^F_ -LZ̤S?Q&ghcaUCڰ_w=w,yޘ+veƚ f8"݄P`Inpzި&0Ŀ7f6MM= !Bϋ)밹9cJ˜@4DUB R"&:^_X{P~~>p)$it[W\(rXa&Rpp+Qw%l0*@sddlȹ< Ɂ,ba66zFxe0[kFh}vv\Ci :i.C>qN|Ҽy(=hCP"`Gwb4 Z-=\l:XHO#oa]5B:i`E+7N~a"x*3,VXc[%Y3v">u8vgwt[EOѹ7JHÎdInVy2( i")))BkVU;8-)jKWgqޜLGHr?UJ}D+:y߂nz5%TAϺ]TLܞ|Q4Aj{Ez׺͕-|…lRz%S3{JPl/VqDe@>,͒ d ܞ!nQ~m ʍ!ЫcXbC:?OJ{9d]FdP &2^dsmV-QP,Z3P^M#<<2c2!%Q։X0tb鄴JSp5!L9](vQD4PFIFR?ּ/3M< 7-1}Ј.X+=4*ϢY0-iқ'7nh"׷RP`Yޥ aǶY&ұLDsE' *ȘJskG-'q4T1E&TGNR87fq5vm/^6ۗ:.:> Y,G]Fw;H$Uw=>1>&4}4㬈Ho*" P&8 6:'D3%A泶H_F̏gQAUǹ.nb?[׾AߺSu ElOֵ%zu%޺FH8 0OG}ksS` &` &M;H١JN!T6bLc*rAD0##H#n={hhG68`68`&XS;G1K83 cÜHHE6YEdaїQ/x'Ogq'Wo4U4q,Z X#Ȃ ۠X #yDAŌQ9buOF{2?z$}BK5Pa:y(R-Rec3yY!s}BÕ0EpUV|mcԇ+_p W4b=` {kPzWW^jEJL$4FG0"xL4NąRLs=]M^_'}Detgw|Pge9wT8\'!X94 +DQ$|̍J}]EȜS0nm'a}j=~}`cj0YsN>&P BXJQ}|{siUi4a&<|-xd.R0٠dޯ8y-DNkKRBaJ'l\wP [ݬamaMztG{@ p:6VRӚC׻$GeϝY]!#^P%6}:Y? ^ͽMm,c#hx_p@j!1 X] =;= !;%n]&[ZEsh8~ihmmH ea,0Fde{ۚmVj i9^]ǑW}oǸ$_r0~hQ ɕ,߯z!ds\ lfOMUtUW[7=|,Hxs!p|.}Ѫk"g{T}ˤ셔nU s3)CqK U>T껋vk6r~ >LI̷1}.>ޣvv>_<7iW\_Q,Xcz51OiD\KSZNyJTrycbIS߾LݟF?UӚY{{$)HIEJ:(IV^YS\, \95L$jQ>n&Z}&0mIh@LBp<zR%+Di/=hLk"ډMaJE>AvMx{tCRQ!@K uXʽHK9# h<|7l`JS8I#xDZDmfPjɇ$(%ړTFA[CA+\*#)7#e.F譴"0`#w!M=X9}FX~obMQ"L6)RQʢ]M>^c<U#iC p07pFNi+驥1O)WqW?"c!QF-r);WI2"u\KǀJ%?F<# Qxr\AV0glL 8"@!DT&Z]] `Iҕ1-wn"e)Hhg߬&HX8`% hK,,=6N 2؃R ;/Ԣ6 ;t dNƾ܇nOqNI,$z.~O'#٬%eݔ"M)Rݔ"M)uSz+DKF4GmOk+8IK{->"d\aMsW_p`ƈ %Q6u,6&Li.׻G~U+{Wl͞X$MX__P5lۡQ5fEᑤ$g;=Ug}nq((S@מבی\*͙f;4ŞOt"є+Ri"+Ri]ז{b1SvL'~Zio^w{v](S[l (Ol]M3ONjo/tc`He#60(k!1aReeJSd2^Y0P9nYLf(({i?kQ4"y㧦۳^Zp%S<@AO򦯶gM+jgy^KWg/ 'E <3"< yi3\6Ej3\6EprBh5 ur|iTzSߌFw}09wtcazy7l ^'X ЯY]-\hFl2'n`Piu*NOۗӏ-KjEu-h{6CZvfijAo߼|D pr<5 ,E*036A`4Xl\L"ecV:JSa@8'"MFr8;| ZOWzah@7ln/2kmRiBʍÎh/!OS hX׋Ѳ>5h} )%4` yi@ .|!bK13P$ 6bNk SeeOcB$RG~~S۾U{Ksdk;zV$9izZ>we4a_|J2_aM.$'uTӰO`#V"ڑjݎmU`M3n7.0]LVEOkakQVp ye rsr# Pd&H]FBg7rNY9ԓ(8ùEBq n 90~?'KuD$тIн"Z_㡦:gHOfc\^\{-9 /' ru‚53 i_L1ut+Ƕ}:hK va`̴qIAO <:/-wIZn>hmr[H}qH"| ՘(:?. '>;VTo'ga9ǜIx?"}ϬyIQOGgeP`')be^Sqbf;` ZuZDt2cWMfO5hp*5N''|_\|)ע֢Z9dވ//:{{D5„XzDz»Vt$#mtU !!^1>K3 M+ثF.kGkъˬirv؀ƮЭJ(AL? [)EF`I%ZM ;\tdЭ[gټccS>){3o9v7ۼyFeê *b9[VNwaQ~d8*)Ԫ`*K, E^hE_UuFZu=2>56XÝPvPDv2EQbt$cHq^QnQ١\sl 7P6v'-!BgBk4i>}XHp3 JLHT! *pBkϒczO'z0aF\\BCh&wCW?>{vtA~zr%y')!ϗUak~ۂSl)F5e:)m: Gb|!23&:ZP;*"p0kX)*֤MfiU*+Č) Fs!&D0htiw̫9|43F{"|߃隣_hRL (AzGĒ5oͷkܷC~7}MpivH~:8;buT+gWei$/7~'¨o!pS|wSUJC؍᣿8+ϾA>t8 -gO?|wj):jW:& re+XY qQ2kYzZV7o+Mm^|̧?PK6R5mE=tݤ[TD^Ms.V$%=1ގ~>*~W:Go+?M \=l:t[ѷ?0̔&G0v3q;7\j :9Ն~N+0nV- JZxdHPrFEEn|x?>mK[V1@fm# $nC ) 5Wβy ~:^~J׿ M^.hl<ԿEgf[k/i#sp9UUX +Aa8$Ji A2%H5AM>ƮVv'>'W"gL;Dtk}s2\)"Xd>x!`NsE!kesK1v;2-= _n1wPϓ|1RoD.XMё XkqԱ&Hq?%ĬA8a}E( 0!#i.  Ix8BcDR"%DMFs9}LOU"p)3 -:/SH/YmJ;M bXƞs`c`LpA w6Ea$I RΠĭ,xn_gGJeҺ֭]Gv( I6>}p[I};ǵ3{hxtArg*`.]D?ѨS`;05k =2ŀ2Nv%URi?=.Jg_rzh~z PxZf٤%'J9@#R̢DﳹbҐh 98Hd8EZyy܇qjñh'3BǧS;%d2",.3& QwB$yj@A{mSDp@i~Q_z%Gc[]uNqwgps?h\d,w+*Y4 H,V~[&1bh`ij:~2sEn%mt\pCy0 ^o(n< rH.*Ldpj4V[%y $$eunkLWr]\n4:Ͽ=y N~V Mf 9pf-'̃d&hϙ5@IB&rHQ.,u!iwk62“(W”8VpM}C_1{g;BlH˪ZvFt{k۫σV5DGǛr4) }9L;""k񗜚%YID}+®L&kMTj7,eUJYY͕O5hjC-Ir^L^;"۩ײ=ԖgJiK<7SkoDi>C'9:J ZjA"WVJNzB HyTQ`31)6=:O&vrzX*3q}񱐠9M/!<"$Gق.ַ4TAh?셏x(W-du`„0ިHY#{ d˔EfylX, cQUtQ7i4G-siv;е}=;ycY4tn2xg* N1PARHd1W^ZnS*@Ti]~dUbr݀B%4 yitg_Arqᰙ <(a%'eډ) *sD#qEf*="˳6>T<Gӝkd@hÅ% ¢AF"}V;4蔖ɐY<|zRgl4! *M(rA94bd2bѝKÕS ƆYiۮ}=yCY6d'q5ǮDE^/YForE8hNP_߃-ɹuYFUZId4iGq\ ?gKӳYDȯf 윛 %<:>U0}P^41&D4+͓k'𱺤*٣f?zxROuv__p1+vgvy7ʋwІ*[ *C: rcxQv^e[1"j#s~2X U,5tqg\yB|V3+]MW綏#_w57Ldz o5F_~x:\'ɸ}{lkh<Ԅ7DkM tV kj'\>! ~QDi8ЬIo|]8;Vkk_lQ([m axUTCVéY|q-҈=g/ɗ7]os>S؎|xB(ON TaI#gWQghJ_75 l\oEպTX߼jahgId[rzIh D|GoYкV]79[BeU=LyhٹĜ~lSH2w)&!C[4~-gJмBs X”(1BP7 \ =d$#5 K)WJNfo|3Eg{fN|pl*(A=`>2!k,*&,hɰWZ0ҙC1x&f[ol"![X8 Omzf ]6OƽTz~X Ű{;e>N{ܼÙs8dkG-L%|Qӟ[t+1R/:8uijTI(|~_5}Ꝟ⧙iY_%_-}@Y5/TRc9=pjv7V3s~:mnwƓh:O_a?VSSjJ8*,{Y^ܗeΎ,zz6^oaD ]"*\Yh hyqN k 7]F<}ϠP0եT]8zO`\.t|YiId']Ί66Ѫ~"'m( 4ؿ)68>DyKTuI0:04wL lw*D %z$yygU7!+N^/aNM\A$f̀ӾZ(rD?@9RkT=Jb&]{ _KjnrѮQ\l;") RNa4!%_Q= y9yA /߯_Lnr/wvaR?7V T8e?+}495{_~nrwZ SIx0BRDwosćy,hoWSw k y}qAx*b̊RS7Ű|3d`-^=nӘ!xlp<9g(=dVA"DVhSGZ Q ˊgY8;~}%sF&O8սͼgJA [QZVXK[87u~'<(Jw>$ulvv7t~'Ԧ_Χ:M'j@J .ŠpC7xBkMt5+`: #M Ry8ᙖ: #5ۨkۙ=ߥ,T%F KDlLJx˽b 1mHWh\MxYμ97N"Ȟ ! 7"HY=~ȂesŶuwQ]`,yɯR".w߱^=&̊Q1b4q8b.P`3"\kjind2I+(`>sgj6ŕ`tGG#==kPi0kgyj&1~:pJ1|&N*M:Ayi:c5^{-"sG~ZO<\`#1`ߢoI0F|}Gk_ֈ:ռ}dcFpϷ.6XGhJ!WӷPڣ1.{[TǦ7鋚9U ۑsX]^CagJ>xu#n̈́ xI%WPp$,X=5z36=Iʘ:`  MdɹdzX̺i=Y#I ^8ۅm E;y^ƪީR43D`0k(`>L\d[jdlZ:*Pd0k 98o,fg !9Jk/ߋR'l1<^v%.)h 0@PHB WOnyni(KBEJIdPMjo{) #%-#n5v}Cf" pvy>K9;1,m=|rء3 [z||\f v봥9N;ecNYO喦)=[=ȃCۜm[|Ll!`[Հ7aW xUۀo<,ypgsKpTa7n'{hze;|[o]?=\0+kY_WkՏ.8q}hּfյQ\A#~Z߯#RkPWρ]=Fg5E .`J 3'f-wMh̵fB79{0c_Aa2F{r{WuD]7&%jpHb8 Hӷ=Xz3Y Mnf0O=H9\k7PJ,uoltƗ_Aaqv4q5E2`S4lC15dHe㾆i]8UX*#3R]`~̻ Ѽ J0,U_l%[Aar0ի;t\P{]<F_Ur>yF,KZεֺpBfCA-F=0zWku*QVJםЗOtOቫ+o7_6t^0#ӓ&dKJ^|6VR$ tf-c96:z9b?G8*Tbl7XBDTA\b,s=N3$QƎ2 XnԪJ5JC*QZ_ x@H62iy%1%UK#ĜmP۫O娲 >jk=t| såZr|}vtlog®M "F'U+Ma8Y8k X#(.޳ԭ.F*6a05f0!HTީo}Ksnf {0!p5Qˣy{7TGrWɏν=6F?;WS<~T"mΌRʶr+/d毡0`!tkICpQFU2K}a_0yg+(nc0dnºn^F% tE"\Ȯ&nmJKPx8buW>yE7!Cdr(^Uwk(`> n՞M^k)V~ M<ݰػopvaj\sae 3Gә'HGM]*\I3z~"ֶ,0 Iϑ̣Ko6M8p(BO&jr $*e*B I0qlnMgᖸ8?=mof txYzVqvYTsuPy6GwdMѸp&@Cs_rd/ Oe]Ƭ=nO:ypa`Tݵ#AңP]($ҙJzrkd]A;$!}h164܆6qj)GI,Jѩ:!e݇xpVv*,Oɼn{ja訄 Ro9$(&IZ+ 9wj,Y)` J@U9C=Xi{$m2XMJڂ[a?gţ #-VTQOpafSse W ٦HU~|t#3aCĐ}Cr/;.Tvj/kTRJF-qz4n*O먆=G/{*k'o$mc+J 7P\V˵)ΗF{`]~8yq7]|89KyPmU6~O|)BزAwO6J}ckQjr*jU 5LY?YtjW!6 ct_k nY'qjH&;.t-Ʈ/첅 ˵PHRȶo&č+AaĤq;h`oui3NW?d67as{ɶ*L= hU`9CRkrnbqf`䥀e 3Uӓj N9xP<^y=Ў.DoڑWE"P n얣9X}3k2.T鬡0y]' PlR2\z: 36̖{l&*ݶ_Aa`,6FwX6zqP 2 OdYQH)Ԁ"'&B>*D PV>d[ՖrAu^R1UY5&0,y07RUXOiFlөucM/ 3~zFtň\2H:t]K(Uu᭡0yp2[ɩԸ;Ni=rh̯0y v::< 7H!w+(`ϮF PWk( S aOf~ ̋ÅT'5VaZbV:f[xk(L`FQjOqd $oTSV&6Yȅ_Caj$FT\Z+[4pSrP¼ OOpv\h)Uל4 ÕN\)Sm!2^]a ?`~㓆5Ty\l`coBC$#á QyPaa̽KW=RW =٫Pr >IC)g˵s|7m8Wqw%F ھq k3b݇/>qm]쩞\B:P_?Pkof\A- ǻ=LW)SE =io9/Lټ!N@ÍA麰D 4d#2aJ"+GNZj)\y$7 뢎eyy;ȃn[YQPy`}-7'7K D[50MDi R䚄`.X"§6$< ~빖_דo|JDqjyjBk(TfpN"jBPˋjùCXƆǰi/T XU\$^z!E&Ou5ȗs0+5''$<#$'y/'>00 M  /_%p:)ɑ3 _'m"Xp0oұ!*{{sj F&ݮS鸏숏 `v~|a^R$w9|+LŹ5nlūW%`Z"m4Z*_Mk,t/@B[]o?s.ō/ՋףY_^v/uv]g׀ ߯靟]ta= 7/ґor0d0D-fQW V0bRbFٹVMz$t 2u„>\~(~\tOpѯp!S8ӫ_suw_ޤo_ׯ~ۏg?}|kH?9Y2/6*)qSVU͌F~u><6 ).̰HLNz `/ΰp.- ^?7_(w-V&CS6ex7qZWSr; Q wΪޤ3ҿYCk :4A}J7.b:9KtQULUُ`(g'a[kSˏ9 !yM*Bʂ˥g3߿ #h>tg) `A0#9u4!` 7S)mmM>|>gZaIdy0NQ#0Sq ψ Gp9$F 0+ j!C<Ԋl@t`1O 3caɦ sJ<KN{TS1WMZ'c+()AHŠm*[_ *'շf+/9W1,x,5d$RIJ?eH(BH7C囬QM!:D#;y[qӃ.{ug;wmɫ>ƥՕ$=S(c ?1ӖT@x $ed"ҳ XZy|GV'`>`9ɑc]k!SPESmRh[HP|?ñ;8]"Sڷ@QYi-bTS@h* M} iRNrGX]3qV;+08J&wVN=Y2g@0Z69c=-&l:5$耢#n(jl(i0vJDB(s L'HnHj`7q$ÒPƙNZ]dJI9ƕĈcEtׇdg[ND[ ;㤏E<3:(+9!2^abave\$=IOewEM]w}jC1ָ/W][_̰RNbۉ}'>]˶ޛu͑%zE)y$y=ȉ"yȉ[-jP3:SQks!+Ei,xEbo(°4\e0]teؕ, ^ZiX@εox39 ;%KY *+  H0 5$RypFk@^IƌsO gю\Y:.nV{y8D8i?<9V3YzXF<~7LNUQQYc|^=5YUE)8GT;/|DKm}!KPj{(fBmWL>Vŧq$VےILKm,ނU'rŎ;\6mF^>-$/UdvaҸ=m1L&z;󳥟"%ьx`}S ?An6[Óulˤ GlE鱄( a.X-.G&((Vz4Q`ZқJb'u@m%Rs@h 3I<..E9Lq*[pnǁ+[/~m8 ~wIXݦ29G+ {f>otn$,3 uJMJI.Yq= X1D .xo+U[N^UiFX' rh'5^ Ղn:P5lOG'|π.V9%B%c̡ ?$^gI ZV&k1Gg-5?|lSZw@e3ki:Eh0JL߉nEB#%G'%X zP6O 1F 9j//X@+QpWI ]k+AY SrYɘť (dTH@}:UQ[톝jyn:$NpRbPl>˱A[. ֐ԗDslPX돯6n%*]}fԵY^M*Ov4ި\ZWl_{\{Ͷ1|=&M(mO7o'Vmfg0Cw_FnwXZ%"ZMS$\2FaTƋϟ3@;n% W#[]q!{\3fol(3Y"V*TQ3~WoZWE&z*x1V Sf}Mg?`%eMLU5Vx/$LQ I__c3g?)NWɻ6JtnqOWKS;k}1\WXM!B?o<)* W 926eyk߈¡Cx9S#gH^ 䊋gTIzNށ!cK 7]4u'dɊO `Ɖ>]0ьk1I`wU07X7V"vl~!}I[8 ~EvޓJ=] JbFt`hs $0]e_NVMmP⮢YӪNG0ZxY'q޾m=e=d8uMk0E+ʃ /?9&M'~ep'wý!l*&6"DJ6 A$Gw4O[_oo~~j,?.Q^uM/\ZJL( 3^GTҠHɴw7>b*HwJ-pjBPRxL#"8%H/%AR !R`Y%OsYO. f#Q !ͅѩD;G+ ^(Aa$-q0qӅcpaQ7LrPRtɰb&.1 8'&JٻƑcW$v/3gOnuV<流Ѻ)K v!j~]U]U]]]O1,1JCS:0eT`BZO o=Q5|TMje-kn 2B[Cwkvv5M@!n-4̺Nw7xAv^}|ofՐeD׽#Huﳻ/ z=Ü$ |<'n"p,E)`iv32_z S5LfgܙӼB,zQI?AA@TH-đ8\VN |[o|*kz%=mM9 [mK-4RAxڼ ^{Y+gрm m\]j 5$R%|H]pƱڄ:R`ڴ'$]Vn, @y,^X%RB.`@ך'aa$*м8!YeScJFL(7419Pb _4Ԓ$Hʠƅ89?ۥ~V,rX'QAђ 2x+(* F-c Itr3<0 7T (Ńb1S3eػ*I^-> H&zpI`GPȥ 8XD)"KM,)91RGDT'O]3Ɣ7Ug 35Fw~|z.60`c"D$,wRPH`@\,q{ÕMb;,aY ,"+PH)+T+ pMZK͹G@i#&+ upuc箮z3Jd"\RQFx<3HO.-dG8`݇sw;#` oj\>xKiI60o!>qt]˄7ҬqV,mQCM/՟af8'ǧGf Til-K\"V8 v` (dɳmZ&tb a̽:ÒewѺ*p`V?b-N֛ޡ6$UHinw)NHK1ūAּ1µ)0D[0ˈ'L[~lj%D Dh̛9RBb̬R`D>M@uu" ʔ"J{D 8eNlKe&ZqkP3&P&PS+ RWXupjBVWWRV]]bV3--Xƨ++eS4UWlh+,IcU6E]!Z}AD٪TWVE}b[l,3MWCeC"`fOIxZ2čոͤ7E\yo3b}JxvMI*:\\% G{GL~IM'g7>ϣ9n;L,pt vҜb4%HZ)i(#֊e_^+ϢW`ۂ)JkV/S/Rq-m_R+qtyw1yޗ0 >tZ ya>NNgP6lIդ% N2'.ʫӹr9uzuY .yvJ[yLW<&󄧪Vl5Oٹ:+M+w8IrϯmN[o3>g6!1+SaǨS \sA}AR(2VMJHڮdԤ^b>PtDPBgy{3K-Y2B>!Jf>]b擴c#?K&/4)Id`)ND}12` ^?4^|=](?K;+.g^s0p ~nnQw9c y[QSmV^T>\!S-X]"r=>ܳڇzhf X8nEK>הB&e#`!!\ɛ!Z}wQv+lspuh=wu(y[a"Օ*#_3'v7oxyùp}ZV.eIh|:;sgL.<[,eL.k,9ض#nR,lƔ[Bj_]e[".|ƖWWy ;.bP $X !w TGY*x+]#_{^AyqP bjq|u;;]7e'M1?a%>/#&WX%3"-k DP K Ы18`~V6f*>ͫO2+]{Vt\+MH!uOi $=>z@Cr}|lvl6F9tkWgF u"%Z'ø~@a~tSJGq&~ґ%/©# 6!J #6zdQ$Yu[+IC*z \kP{d!n~1L љ<Z}ǽI_w +p8_,{]|FO#}?`)lKYd{O"Aofq]~jI7H#@Ciw/AHʬ=V׸)'\QOZOq} Q&ȿ62P3ıV[8L.fɖ̍u袢x;b Ջ0o>KBp#iY0Hp];ZG0tAN+991icmZF(hs)[bdY,bdyֿ0tC:ter5郮]~$+>.>/qoI*ì)XEjE @$pƅ*+"(? l>e47?76b(W̪/|;Ekd͒PV0ٙ>K^cS*ޘ&~p6tDl6JqRZ.ci2jGμN12nŴ9mF /1[!Z{"I {e <+XgFi環8e( oU*eXV 9K+$-X!K H&X!R;j)dH+PWN4f\$)30,47x"xiT /I(Ex[IL D^a~|7b[C2Z޳t#=3Ɇcq?nk ԫ6]X)tKtۅoir64_aE3 l3FqAތdKXZY(C1j%5;%SP Ug},Ҷk-bѕ!8 ixLTB;!R hHCq`ةRscX$)챦H@7; 4՚V&5 i0ZLIۚ|ї$Mu.X1}`IxJ;eKoeM[y+ ]Ïr;Ԯ[{ʃ?VR"GM qX=j'qnh[mJ,p/[;.sЕyD۫ Ey?]B`rZ8䓧 vFTMsLI(bc?-u6P>͹rƣvl,p =f_D+Y ;# p++P'+Y^Y |{L~}~ F/CQ|)L>[%;YYjj~-j\z׿m~|2' ѥiuGVpEEʇ: &Yit^]d;Օ j.rTKW<<"yRzĔvɕzrh'y´09^L쾩A .Y%EGs 47+0 SdI dVpI{ $%c2cPY7&VrkSI[ez -äA6XC߬ŵ 8w yc g\%#&%Ry*TKLyMZ!LV"Mšaro/)nRj!'\]=`VAɫmeQ mm|2j[K~}+ΘJ[ 7rj,$9Zri{M g^`Sm`v`Xj $RJ[{,s#dQ :*J]*]m0AKO:N)sL09 ͱpJ)r(X_(1eEshQ\~DZlq,Ct۱I? k&q^?`n~&/6NGKe{jAs.)gHۘSQvͳIEaSׇk}2 ϴҿYlӖObv=2G™TaT$MLͩ,!Ƥyf$4 (48Ueҡ@<,+J?K\J)ۓPؼ<ߏ;_&?'_b~k~(O-ߕ|xj'z%mdsXkt%4)RI& ;׭qgE-5m8:sLFH0@@XW4dEɱ53im [R#;KV枰te dYn{>k+b=A3טs,tu$K#YZ #ב,G%`ב,=%HƊRvr2#Y9vPWVϼywKdݰH#gfP5 \gWSv2HQ1 b:EEG"1>Jp-tu FYϑhIvڇ#1 Q+ZAp"XZ:HHe5KV}|68Myv:3+nS2< FtѕLaC evҔ7 sF#պ~JhH [߆ơy>jÉXrLT!ƳcB\))r NӧJq0B0׬coaTcX$,h$\?p|d@VM#8+q) 8xYz,jPDu&|Tʓ(TFwK–u)ןɷAsHd2%pRGlVQj}QH(tD^RcE)BOm_`w9`MA?iq;zNXZPTX/7{"A&;-.t&s{ am0a,F@i6a#ìG+B9p} DtDAqH"Rx=lq+BВ9hxȞ0aǕ@A).h@7s&KjfH }@wwS=$R)>hZ߬A˫tFwRɯAh\-$ti =J/,qWBd]fBң_IТ׭mxUnMW`Wo2?PG8K݈/?%RykYL#lKWz(gr<޵a|φw\+Jvꭢ/KY ZϹ4K9G &Au4# /A?P0/ϣMe$~Qq7LV!4"ר Bs/ 3z|'y IuIL/1Ob^yr(=yEű:J}jB=d#mZ!as[7?VͰz[rG'L7 W}EͬIݼ@wc"Ҭ B߿!L&-}0+ҦSG'ÆjEXx} sԵZ[Yu%q\"phyFQf躺qO*(ݗoە~kc'E/ )wAY/nP8z!6 @HabǎRa>9ڿ.`ٍS8<mM+-#)Ѳݗ0D/ZX6Hua, c{. M^#xȌIޏ!"VyѬ(S5׼lW3jG:dzH"a9ty}h(xUot.|}r;~Z^]Ͱ=9#ʲ}5_+J\t{Y|n&hؕ+Bz6>ugss9zyNX@1'2ҋ>iߛߙvSy;EPEz93L˗wb0B*\P_Vy< qɋu46oo$yj~y$q{H{ ̨n _g%ó֭,tH t-;,÷Y~\ͤ4;>K=$ 7EeWպ5\!K O'!9))e=%.t\_Fc$ -F$ +֭<vqw^BS%,S8з4x }GG9\B[j8Yk7qy^c #;H 6Tɒ(y /kTh%,ZĠCdu s48бyVIƅsRnvGWZ-xV<[!@זּwmdz Qg׿_ /| r׭}ݐ<4 !Z] y" k\U}8cZm՟rwf< ,+LjqlD:^s6mZ5tsuxˑJBфLn:e2\AAkv\z'mENbvmeGy-úQ*Q>ET J 4~3O_ V}kĂKu</A?E=ޯ^,E u6iMN+rj,t8 `d]Lܝ}]**|^BAE(qh:6V}#n<%a8@ JrX]R5cG§ϻߗޓN]{T$IM"!(C{p)'v }m>2mn0vÍF%Is e`!o :>Ex ,,t]뤳X?X=*L,vGʨĐDaF|^۬OYi4Ng(:$ΉG/.Tf(I C2H2)5ckz*ٲH<_Gkgb]v7)X:N5!:%$W:>so)?^IwG]͂rѡ36LO 0d$#y O3P;`_FPMx~tJ7/r)a 0-GSw%=f0LnI`2dVk[]UmE$K 0Hwu7W {?ߗay<-˟[k P 4Һn}]ҥ#42OeQ]I?%# LJB3%bv[(C1cG݉@r/*?e.U ~#iWҀcž-5/*h?kx_C nksє݊q[ v0M9s57i;鉽@ɠܱ]{|/[&[+qY&(y{z5:5F(,yr_Հ}uH}%Q Q6}s7Ef)-d1pĴ0hS/M@(, B)0V&4lA9y+V_tWaդspX8!t.0 9KOBڭأdž/ H{Iߟ>B|LQ+،|g.ށx)¬~טim!/i篧]#3gF0U$tH]0W.S#ɨ_ǢNPE_䛙948*XK0jPQXwxM)0%nCcpr޿NJXBFQ~侧Y Y|ޯM4NPEh{c}?*&cQ+ ^֋>: )J"@2{LSPǧg|$M fŠz.V`!mH3aSo|ȇs|7 SԑH%rlQD!c6)cx P/PT{,} #7YzPMyqLN8}90qm b?۽.㿯V_~]zJsQG7gR8/wW @IFz݊ˌ˂"lo:gZS{ iB@c L&lm(Vնnc튦ζEhCs>vUQ82H%P@߶|cA@Ebc݀$Bl&be6"a6RW\ H\]. Z ?-e6JQk/o"8*J=nbcObUɔRv ەP_TGO9O~^m1mT t=B8FNV(L?$Of~1=%_|qW!(t֑8ؗ`l5Dߵzܩ̭z9U0u(DS.l. wU?. BT!Etoo_Tc_Vݡk]WQ+b+Ѹ\bw;?jv'mfbGuHe;R Y}>1h6O/[U\/Jrp%L u|ZoiC}세+ ?fP1utNKޅ u?<1"e?[[RoaXQ8D;Yy3%8NT A9Y<A\jmh+[Yq;Kݛ<,a +yDmpMbXN C(}zC}oř uےr (u!"ke[(JQmaEp>_=ϧ9c<+~mAHX}}=QOwbiU-z)q2]uMpLF t= D&}rKR:;^ ݱQ.Ѳ#NbjƑVe0pwZYv6Tɠ@.m=jb0q)j\XE{JvB.O [Si]:e7Aq" W\1RŃ,yEhE݅z+d펰2z/ªO2rnY5e!|[#gUǘu r X/֐XF=Zj8쬳`L,-I⚸QfDi\}`[s]Q*yO)ɽ=5kDd/KDOJꂹ`&};毫3?<CKs;{yn(2q $qEVS"1]r͗Wz= JOf~6;eMKp=j5}VSΖ Y% DrdBEgBp'K1gɖ5:\:4Jh3PrZ A1‘,>ɖ׶BMҺey]o߯yYY}e`E ѥ2̕l#;6C_QgܦFRꔿ K>!To^"HQŠ=ԭ{[$5{L`Ȼ2ӝ20"^Bitul!$ywZG9N6{b5c|E{ gv\B39M g^Wն/ 1 ̫1O5Vl*x9'bb.i{ ӫin783J>ZgX`_-| 0R8K&+FNM{uOfiUt'kCEțAts*đ+84V|q8ǺA݄Zt8jl zْ3ϧC Mo8vݰ A"L@ۘYG>戍#_\P]}bu4>c3@,^xEIN0z*+xdZ{Љǡ(wq ĘrcqnwmuBCDl<(d:tďmG5yqx;ktx9c_ nm[9c0 Hfey1Ne$ؠ> CR|JR~Up`]cd=!i1C>roC$@3( h߀4Zo`(DuSiQ_TCb @6Ow _u\XQVo(vgHf?+ߞw1U,NލEDc-]S<6cIpc%(תj+XG6 apG#_ s900yʽsZZ7:Ay 8G|"Qjmp\'[EW)q(Zr7N $MW}@0[/i(QPl(̆ c~1Sy7&˿._=ngrvT?_tf\eAeuB{K Fp*Q[%Ο#dyBs;CBsr`q/4S0lnI ;wz|In fY1-_%%Mn`zW=,b7x=ŽZh ,ݪ_8I &؝Xv:Ųe^vrBm4TqtO%D}n>߭Joί v47f-a CO.P ڨ;ŷZfGnUsh}WҊ"$ :DQ٬6V"(/bBJrPqk}1M'6 _7N>Y0{V,$(ʐEBÊsw&'y _Kޯ#\_A{4#\IDOӏ0ED+ۥQӂ$<1I% Ȣ6(j3~7VQƝgM uy !ft8B=q@b! DD,ZV8,2W ,NbI/zM-67>L3z5 $ɀ  L:F3,Tq2-:hI!` ,W ~/M-8V$Yc4190Pz-J" BP 1(h= IшqIol3 vJ8}`;_P3PJArxd#o`%-Zz4!b, G\Hǀ'TI QPyXPɢD obFho&ŏOw@F194~NCvYD(zvR~>4Kx~Aj꟨-zkfEH1Z*f8Cl3p=-z}x\3V $arDP(YG#DR.ӌ=Jת㉺LDOQׇ0EmƧAWPZ}:9A\ֽ;K6e.mbgI{H+fU!@X쎽AL2_kT!ҤH: }_5IIvzXU^WwHsn^)% i2_eɄ `.i8~M0XM I.-<\]CDˑE"F%%J|DN8̴Rz{LxF~FPR%#!ҼU`,:'~/Al0A2I[ὔR.ҊLYB PqM4M%aSo7_7wED/a?qͿ^%OONN'Rf T],ź m(g)VamӨ$٩'*Bu|lnG ~n91dTq(%iFk0 ?b`=wQ}|nešr]\B${~o€GLj Il çF;S6FQq͵mATHBH}̆/eކT#nv0Xmx2a/w!G?8jk:T QϥKkC[KcM4Ӊ}(Qͱ)U7Ӟ 2t䏓d|5Zt jB}5!U凔x!%^~Hʉ吭JaBaz5%ZP-7[26 prm Qȶ)/g| :h:FԚRJpb&ޯa^lZ {R!Zh yX\׮z>+8@Sd|{ӆN&y>?conS՛οFA~+0,,(E-nr4k;/6;A`p[~^uiʺDdæyJexZ_ن&ш3u\W=gQV'۟\zm$, Oon*\*WOf[,*JrPm4q 3}`bQIe`59c^—{MFg g/KMS ϗK &TI:>q(3 i&dćcZe&sL8p ,Xm.T$y{GZHʑVzNU,E#Hb&q$ [@ hƅ .H111\o r=#>d<35s;9ix_ޑ6krTSrǶ(pH +i m*eo,}ueQ݀ b9"t2-:nxn~#\ /{A;Q8 Rw_ىAZuS^w< ܌EESս(V Aqs#|ܞhTxžuI&00>D)toC/0n4aT흞Ϧp- }Q:r+!@;z)hlpCos[.U`,e# #m ^F6~Wb5J5BRk~Υ~ϟcPkCΟ5BMXяf9qr"FTXJEy%"}sڦpoxJtT}٢kAqp $ØT Z3ssz֗sNdѠiv`hװ0ke6Dg'5APaØlkrӜѶO|@y''0n7Xw.kvAji47JJjIZsy>y1!)UC{jnT#ً\QgkLNihВ1״F  ڗ;WEG)FxWO{`|sZk:Mh9u,'Ԩi\e4ٽ: N}ږz>V^{`\¬:~A#цps &T-X78EqW;wY>wd*(cҠ̤ɣH^YX݁Lk%ܰU`d "ր8J6qn ã4Ya+@+3h%m7BĠ~eׯC&Q[bNuzjIyB"{fGY93\Sz6$G`|ҍ1#`4};uPk!M徝Qo U$ ETe *Epyw6s"up֡\sBÑ׽~`,vE*IHmqw'q99#uH@ rJ⟇3`A Ӵs,VT3F0 kzZgF75)Qhrj,fTM- iNP N`5p$nYQ* ;̮=᜘|P,9k#cjY|AF7_rXK6$hr! @N|p{8m&ϋL~C"?l?}"Z3B6/ږK,^|N1+d3M<q/'R}j`],*PvԮ@oFՖ)Opx某̮ Q`P,_xYTI4-|-5 #$S]Na7F~X=8QwyO[w#"hOa:兘EBgR*>O/g WYFddVCd~{tjrj+{tm=zKum؉yRpDPBe$HrW!9"V$Ne/^yiSWh1W&]*bu]7ȹ1ern^IR^$bEEm&a8Ggee'*wGր+mFэKe# b%Q!<= )||%f+Eݰ.Ne3["`9-_=0/?[&` FLh UժVjQKܺ eKy8Hu +"nK);D7`*@r}b3n4RoˀZ&: TL`n`pCGG8K`DhrWt7C-VX;4w ~3b22\թ(M ^HhH(G0 +!ѫWZ7LFiytkP9TzԥaeE}*VK_ʨ@a6ZTA]#9._l zpƅ^.E|xreaH":}A~X)Vx5;|Ng'5nTzK;VRh:^&} %Hr R K;~D={=B֟L'l{Y##<[~uU.ekH %*n-|飪wޚz%O⵷~΃&<93Ubf@.ܗ]Lخdl%l_>_EGWRfT:(7H1E8 VƊZ%|FFCvqH((6 e0K%kd˷J ;fLR=eDNEB'E3l$>惢W`ovM.3fL^ՌU/ʘD382Uw3v U#X%2R;9+/Tvфn{>U$m%EŨG|V]y^M=0W[M Vr ^S g5FTy[~TȸQ9=0.<`t92u r:y{gI|N3 KVƚ3BQoaj6Aex:=Poop6{Q$\_̻̮GXɀ# :c{<>֘Wi~h2tb}|Y{`\b{%t2Af֋_f Pi,r_׫҃^S4zRt1{]}T h)؂_@r͛_&l( ObԣJ*ccŘ$?`M;Z*fu ߽Kj7roC<*3tw-oUnʊ+iXj7=]ǔOWC|v6Nv~O81ԋQE`NcJ{F_eg0#@,vṅEbb]d`V2-VQdQv 4b+Jd=u3V/_t4 njQM,n#;#ϺhI˒RzyAq^!ɃܸC .WiY {ZHk]eeq֊Y̌6tɍ^JDF.F so5H|ଐ+~T\U5!O M(@QMh+tlأ8+p{p8]o_*& }P؍S4–Qk٬G-n]*GGKУ.Ig?uۑ@Hh1*]0tN0JBQmLpqN\8u<rg n6=+R7^ Gm 4P\oj9x OZ(B>"x jsAAokÂHzO'}|;gTl?3@e,#@E [?OWw3@/Wdng+ge!+6[ҷ h7+׮^O3zxҶ Oz`m5Ru'q8>&\I(Oi'Ƃtl6PL`dI}#'\!# 3L6OנY{0vp-L+w+q.tܝ'Q@CPJ=d)> $`&{\ĹŁ׉LǍuPD-~N]bzyNr ʜ3UH@sik#\!;&˓2ڮZǻQ1n#^γ愨F(J/bBF\1x0+m΀'$}L໭Bf$gmS s8IIT\ $7\@ʹ0# p_i2T|TT)=VޯT\"/+xz!?Hx<؏Gc0m}mx+B  W)Jb vk?d^VGCGkPT+<"3vg?'mf)?7?Hڳ*+ xͤ؎{S&N1G\MCyA} Fj)؊ԪDuY@qv_ߏk31ΔDay40FQeRʨ*}|Djmj -}_>bV>C#`EoE˸>Zmjh- &y! x")IBW32 rY4nUin񤀦WU=z'6BKsL,g: swiRO|)#NRPΏuwLuzoO%Ȗ91d {PNjoqsY5#)\~qnyjb#XGh22^ƙ%!˕ ^u32La:T$`~!Dy_ 1qUx2$zJOA]ߊd+!D8$1L3(P ԵLF4#L-s&CwHF5f~Oz-7i%u,٤1zH+<6ڳq+*VMs*p) ;dj_^:=Kۻ4g*@HvyUqWT]o-7B J1^en:o=q<" Q{Es]T/EE3bD.j/(BL±g :%h3+A ԑe/棢B#JYF+iEw0p,FJqJBQ4 4'9'vg4(SJF[G/%c!#:7݉} WDpa{R-q0p\n,l9IPث++Űx[TWpL>5zS!29ܪT$03"IQB ( f4?:\7Hjý5Y(;{1r4^k^c'yx2">o156I2+"n|fU 1z&TX3΄ҠBIͭnmv/zs4p3X,w8 EG ttg~ j[M܋f|(Yv,r*Tcv_Ⓠq,ɑ,pܭ]9&DP{##I0HJv2aB?ᗅȕpnozx3C2)Lի>X!BPq;L/~?;S o7f}:mK/hc)-c| d]IQg11RZc 'HdnȰH ߊͺN%Uq!r^ ѝgu?gěJэ9wKzyTȸb,px'ϾzL|b/'FYWtXW>-*d\C# o5Tk8]4X? q* 9#0USQ^2"O/%V%RXsh4倘H(~3,ϋZT .ݙUqBx^>a#XGqؼ]6~G(jзmU#P?xy8{t,ӧ5nd2fٳ%,ѵz5:k*lHxXq\(JV4z#e u([n!Ů|9`1}OpyvT|DRΔ*]~W1(\?,<Çg-ʊO@!eВ3\KA0n`HtIg[-Pу$^E$1׽:ކF^w1v-9[AcybT`;V6HP>3NqkD9'VRԪx?E#h YuT~v&KTüui8Y{gі "$P%v\>}= <;zVSApfo߲-{d{lo+-\s@9^&~d|%fdck 'zN+S4^0|F<Fgz;R)h%L*ˮeVxZ/ Aׁ*m1M^q<ꪥ8DzVȸ.F]o:l8_PH`Cq]«nYGz:h)5x}xs}%^+ؼ/8_y+GAB13R2(`|QW3:7cwnЫ&RF^hxš ] ]P&:FtDd2ک9%@9;G\1 *dTzc4W1LG}#ɱWpTF;  ^!X ""3);: dxJi8ct1ʯ/FsTgscӖ*dļkxg>|f5:/J Za@Ba຺=.] ̏o,'|8sXjsMyGWȸBZP*\1l)c4DspkD5ϧ x}.u,偮Z\~QC*չpb I&ȰH d͊N%Uq![r^igGw +I- aWwKD*d\oZN 5l,(}[g1\w f{9t~6ewC{)եnXK1`|~ >ׂ(P383=Mg{8 4QabF5ݯxifpc3G ߵȗaD w'Ҝ}vV}({٣߇Eket5ygT>rv~~^~o嚇7}$v?q%]` hH}z }b# Sf<국:(OoR7/d䴋{:q+ڿ-8a nfg{OmA5)].V5rW|+d\38#ADi%s6 R gNtehx}PJxK((ViOM,}"QL(jWMqm>{1ƾph٧L2P殛LfcVpÍ=]Vf8P&@ +"ƭ[`~-gτTl<y^ЍFXqA XVJ,(0 gk1 ijd,DDqP!cUAz5R%̪x3,D΍F g۴bBݾ/oX*SCZVB5^x D֔uD k }ߗ"zVA5o489=wJ {&?u}}5 0}Ȫ_~@ʿY$OCOO-Kȴ·S0eV8WV'D4I"CRTۥ(poe0O+ӦhוTw'CWTЧ5Bj:5l--M&\\< ] wi, lA=eWR4}/Wbo`ւb U#;E[팱$1Zd쀝652+pG>&*Tmնe}n3Ԟ'ܮ?:g^Nӌ<ͱԴwjL íRݽ),|Ĕk//DܜU&E'mVG0OWLQ % Cϟ3^|gЧVwmqzk^Un "{^e[X)%{eYjCV,UW,V]_7]ub-X&\Y{?yye|+eϴ_9.Ki?q~}K$@r^yaO28=zkwD== mb2,sNs|q9M ) XT>[%~y}!fre\Alm ޑu2 kDpEA,珣FnbEdg;m;j;/Vv[D+5+z,N]A7qS᠂lԏY={Wz}ovER-vbrO?bUEWb&S"kn|#cy[Uyu[<_omO>bF|]5^H^;g;<ij1$x$sZ˭]ߥeH&~Рe8 Ts+F:\1Ekn5^og&: >X S@Mx16CYMŏ[lSDoNB!6CEKN=]鸥|ߝFxo?6|atcM`ڇA'x <$(rCO)8 SJgY)4u)lql0(uC6>kcMFgiC-C27,@v1eFlKwS'qGgjQ\NAmNsb(Л5໏Gs?d@;iи*nƦG̒^M__-TL'xM*H%`EYuz3A`9Sgm]> )6}zH=POQS7m55lalX;ut$DL{^zx+:x.7K;cDdѮVU-T<b^bіڒ_yyYsbrwfI+{UR&8il6uƵ'3Ơ{^Y;cΧc֞U4yLیGg-n>)|ALb ֣Gc*Er}F>-|[<@π|ɭyV;~&3WY'#BG_S/s}uLJYm }ƶmm KU/j \ݓ>~~?0T.XA.\;`WwU~)Ut_Zy,|"7UX!]*U}4E:!{uL3.IzQ#[U7b~!T!㷗Gfe]eODRAAi!3,墨 :eKbu$V_SV[-. ɕeC*|P(U.TL S-.1M$zOUctqVQg9QcH"~R-mԩR[4]˿8d-]AKջ`X_ 9V{&-(xj0¼8-w&|EK"VM[= @a}"RrMĺVZVɿz~r?y;&Rb}EBk Y$2H>YQN-0% 'CUmPhdmlLa_@i,)ɒ]*^B%YC*1>$6Y1}y]%~)y~~{)4R.C|低YH:DpI ! P."k@#zAu`{Y(]Wu ٹD1HQ Sv`f=Qzl[?$?$Tj ;1Ow>]&C$0vR(1x0nuTiPj[XY ,Yu.`]VY5b=G!\AHcmkݑ,н35c6NvNZAe( z"4hoPTy Bjװ g8*i'r!H@oCNEVh.]iB,(U2,e~6h JhV r *b9c0]EEd#v +3.MIS%M#[L&:%M1WK})\Qs?ǺE{'xxna(Ƚz?9貎P?T.o4޻ +aʸ9kGyýxb>-e !]}Քoq RWO^dМo{ /b~}b{Ϝ/_SoND Ww6z%e qehp|,JՁ| ~\8_7篇λ+ׇSyupoWS b&Yk>߆Zkkxp a`FE~v=># ^0F-ˠ  JqQ6#qn(Ro/391%(P1 R>΁ h_#pzIt5`+~E 648񝭟0+6|²뎔Am-@-=jP+ΫyNZӼI:BWF?O";W h1 j- _@ s;'_ylH5`yD#Xn-$s<'iAں5n!-)4f= gomȶf{Ma5{J<5[y3 vۏޏ0[Vvv<W3ͮ"+ntm =kԛʏUq]{g?o}i w(H}1R+6iDs>ާ1lvrW#1?(#)N!6 Lr&~y\.x"=&$׼x-J/YHO^bƼa% @X=ne* %iY]iʥZtd T8q&!FOólh\vݑ:zG: ET,]dlTUB瑜 -evDkN22J"BoM7(9#:?ݍ(;3笚f2 4pm>HX3Om\M_tDr;o =3O{ja:@Ȳݖ{Hr;k|"j~SYyQF qM<7M({c`??phi,yzؠ %Vv*[R%a zg=R^ߩ6XumTmK-u8D`cY ާRjh⌑Dj}F)S?S4N־w5so}Gw)8okۍIb=FD 0طE6"mukm~Ou)J2K,~pKEFFF<@#2w{xcuľ{z1DN=i䫆kj4\Svõ9IbiI;KlN燳_řJggo?P\M4c`;&01bړ攗ck뮦*6͇({ڴlz}MK/"dzNCĨ=<o:j%dQ˳sZ]" c98*TS/rb0I`qĨ9k,BxۤWejQƯF.majǤ6w"m+x5G zVGl/p+:ѵB9.vBl )/aJ-Y*nʱAm|qjmvL-zAt2'$}} LDӤXc9?憯k<>#!iNJǗr|lXbţ|^.&$g*TԆO,Hj(1:F©Y 1:4c{7zڝڅ[bsN2d"ƞ6),i҄M iTE\ضL}9w[A:P09R#El5.x2~n97m,un@p rEE@/:nz*r2CҢI_nq<&c 77yeaTGo wzޔz]t6݀;Wb+ۓ.}x_ɿ^\{Ϭg^`d?}JgO.ʻ'ӿW+?U8o}Y\Too^.1R,1#7e3|Ǎw< #_|9/ XQzt2NG;~r>,y'>0֛ߠ ≭!#?9:ZZ}pO ڑ="{mj.oWJɱB5i朂ΠDN7Zbgɀ)Y8&1P FGgOW|#qf a RkU4RUΩ5s,(Ƀ.b.ڏ#`8\h1UYg P3|!CYȖ ʯ-`9=>dy} 6{/|qkx8 vw;g)^>|yHXE7p7E1KOF&1O/QX7-ީLȿYo[s\R%qIJW`F$ 5I ^GQI Ӓ-xekK=XߧoΦliv$@ʳ֯-Yu4t4A0El Yaag5(GD;>;~W7&ˣAOӅi[wJQ%2f0gZv> #ƀZ`l.(})fl1֚_͠u$NJB 7QLq9tu8ay(0{Cyٮzv3[gr^sQ"<zoz`m\< Ϭ7E5-1ef"]Y>jUKӣ&)k-Sl{>eip@%چp氖!}X)3EQKU\-½Ȭh+Z` >\q[ku˂t%k$52vB{d&W7w= JNNt!:^2&;ڥTgZ%@Y+5&bFi)ңڞthl|ué6ߖ8wۂG6Tz9fW Z,!sڢM͏yk!0>ÂwU:)c4/rڈmkd8Hۣ<"qk^+{0>5v(+y/-`#gWzR- pMbr5CR"G6O6@*0 ro#|7$g8⣼;N ; *9[jG(oI×* yq/ҍn\!Ħs`fE[:>6s|D ^ J5%hpjNcV/1gSk Ӎ?0^&G-ҼE5mȏM`v m$=ᘲS3v!pBvK$~0*l ^.꡷Z˳^%3ފb:Գ6ŹB@_@Pl,b$5)9p8S/"p,ǧwC릀?(!0=*_tQ[唎!4$Dqb_9cfp_؆s6A\FvJ9VnCv 5|JW5- <Ǿ[|r2UBS%"Rw׌͙YeLv h ?jƆ\ċh\ȹ'.}m/JU{O{w䳊V(o8K X룲Cui]D8 [أVĔ |[KX,-2mWxKNs ( #üܻ ehjX {ΡXac_".#PΈS^6ViǺnA}vgO< ]g n48Xw; }d#,:(qAOk_Fknn,ĹZ2Bd]A4Ze^5Z'9Ӎ|7g:q1x]0]L7O~\ߛ;isޙ;9 w;'AF f@K r'@ 3{ 7;Lkw}ίdWib݈|S][rbaPNq!;y Ys~ )YaKLٗ52' ؙul]*'qe0͖2hF29N&HI}HgkKꙡBzi" A)q4dgJy}ίab>opP0s[ܽ&iL\a_7_D׉.Eh3lMԲC=܉,Cv<; 黙}3$z]ӿ=F"gv1<9Fk=BڟsAc~\0zm IGN؂v]3G>ޚ2+Ái"Qnn{ּu1HfEv+\k?׿6W-ת%hv?p+"b nϊڭnO| v~B١M}x0yLڭnvh%lwZg⢄HUGIٯMV͌8N5Z.x s6=ϲ}?ܸ//wq ?~|u9(z9N-o t/~ =Γ<5"fe>ӊ.+v:F Jn]ʰ!6ZJ/t_=JN'Æ M)0FvD=G˰o1NIuR7#&.}ޝl6]:aunRLUl΀i ;ѹ4gPKKM"@Y7)٫)MZ#RmDH'HNڒ: FU:skL; sP~0v!>dGߕ5&Ƅw>!oP}s Ou"uNuf+>6 l=9*,cTI;pWq]ۯ<i.ES\AxWǵ$'DI,U\wiƏoNO.YzɅQ_<[dnRUWah8.X:3yWxBtǨ)WhރҌmcfn6䴄zD=micR+qOvܴ\d`G00"9?%&'THhEelsx%z(*6v.1=uu _dO_<ӗ8J3MX*qŠZsm aqip8 r4/2Tsl98pSX~4 [a{]o0Rɽ``?-R[ eDMQM6=2+uˎѹ8z dEt-F*(fjD@bR|j;ܧh<-ҋ/MB 5m92FSYH JfC6"s!@<.KPǣvFMyG*s%%rdYAf(K&̮Iy(XfRWHj- ]6QXGT%Qm*}$I2` %#,eт- 2?T {$F5ZE >7} 7ŕ}iH%Du8^ V7e޼8Tߜ_o-,ʣ$Ics5SBN(OpκQ:6&&euym\-)xKø䝪-XGzk<NJʪůcDK\'06y F03#XDjʍ R5` *vN)|v TIaD:_3)/x3!.CLi#[qWzqZ#h/:MLp <={Y$o_W1 .P-8yƗ×uuÑWsh~o2HQv%& 73؝ 5κ=X$K$DHjC.0#^k1R>&Іld D_ |<"o_{3T.-#K)>zD4˥n'}1D[df,Фgy~˺D\X)Uce y]Fъ5 XJwǏb+P2cjEC-r+(;\\Om>t|W?(*,=W,z np2ltLwӫ0Ygu ;)&mT*oZKp]ẇw~`>: AԓDb 4}ㆿ LNYct1y||w32=)wpqGd82DEBþFsz(+ =7IrQޭI._n"2'!Z05ero}oTk(p7;Cx'Dr_I0aXqwZSh D9O?ҩ\[<_<{Wu{ֆ}?{̉;Oϝ{]`ŷ'J񴟖Aҟj3R>kvlm|QYqìJSx^qT(c;xt"6<uF陉VY{FĈ.P1?z&2Fc'[}c|usKyXo;M߀cfr~x:BwGvHٛLeёU Y'$냭GQ.m˹; +2d=*^'2(+ M~Bq40|F~3+ar>|F}. wEӃrk{=ddy:<l:~gw|(}O~[PrCP x+q6TBB[ T*(8[y%m\*Cۖ:p]W/+iĪw{9GD8}m0]s;:G!hGo5 t"c9|hJőێyFZci?%o\9˭ri]~J %J0ߚ2?% }mKmpJ ~'EBd >nߟ 72\b5>w^R5Ԇejt cU:1sss/ũ착<--sPܥQfM]GYe" Atޒ>FĭYO55~2kcڊ ,}9hIŽYbx$5k'n'IM.ou,88bÝ^|CYJƚK7'7ueGwkRhGW۸fg5Y3P60tDh95G2V"WrtA]w^dGaqLtǍ*vxDBS]J8ԙw-lHrbNv V`&KmV.YFmv*#5S/ufcj t7+525j6 kbHK*EhĂhKƮ%>otd7GLw& UJ.ĹAʚ+± C%ߖ:T8HA8BO нzHTwVʱ];K^y2{w0:N촑6{kIޞP;>lgOwiFu;՘aF."LdJ#[YaגSo.dAWZOoJ#R9y2=>g۱sɶ{ÈR؃g95,b5PIOزT3ՠLQPPyճGtmX; ʨeGs~>8JF7` -$N1SP=WGY=Ԃ@-"Bwz:FȧqG7yW<010_J?v룡I]I E#vzUxEdVf<& =Ur隔4ub-- Z6,.t\S-T'j3鹵"lagzܢW k}s<`Q4Ί;+_y:b&+`7=p̭NZ{kU%Ä$UVEZۚ{d;肆 ۱rz*[/44Ɩ,ɡȸ!+ :^>WJ#hgLp QL$TP9 A/zȞE^SlSmRLuKQLE,Fڔ=bp K=Dl%)@/UvP!hjbWD!&ڀ2[Yn[uqRv^ dY4 }˱1JD.fn?YVVL"<3_Ϛ:rBvvPA$oC R7K*{MKzM՜ʴ.!A5 bym^5& baY\PTDeu7DS'##R-?ƭ!)T-uGOu̟j|[}wlp?Hrz@1!`MKvttúwA;ZÐw#;Zl#,62UF(FXgQىOS~d;=E` .+S|CF{X!s+3VaX-`B)d5恺$udztO}K~gvېcWb6D 1v7%\uV08y z5ţGV%s`gZ%1.}alM@2V@lmϛMD) nbOmaZDH8 c"8;Nx!!k[-.ҵp;$B*1|(#OTogõO'jHE*΢v1C]LJL.QN|#Di`Cv£y$Yx4QvquvoKD{s$R[.Z%joDJnϩv^fLh1ϋZ{H.QHnG],ت]#8"ӑ-ـM.Q|sInN#ԸNذ 5/Tc3QUDZ{[k{8%/'*_gFŷ:$U2ho^o:)T"v'Lu [:$Ş0O J52lWg:ɂuw4S_2X|r*?<OO55ymME( fSY(|}5|6]=℟~nz+;5jX 6pS$7*՚  M V[T7ZeA+lp7"eqp{[;>4ɠ7OO1ŋ|򩺹ˆ"8VIŮ; ,|wDfjOO:1N wL]"Q>>)>.*ox1i,>93=418To5q8+MD8F씗lgx|W\9Yf9|^Uxa-d+)6*|6G뼦_EY,~lt9}Rܹ+Q :FXVcvhT!V! lUؐZ4ɏgS{("7[jOd|:0bbaa q0TbQ]\=moTrCb$uHw 2RCNr$Rb7Lu{n:85c:`Xo&;Z*i\36獢'zp6kwV1^ NlIq(d2<`ăoA]uȨQ/J[jᅲ]NJ,fd-1_KUanXLq,@{%[i>`lPk粿?/ڄj2w6+dPʏY\q(lMC]k'ƾ2E0u^ ]'19hrE5sST* QdB*xqFS/]{B?7H%V Y47A0ş\L1fGcH64 ՗/y%)iAtoǧ6C+2*>|86%֯<zIp┨Dn4t'w57MVq{7JO82~T\hg/u+uU:Fm;yRo>wxYG]mQ]?gu)BuyGcYdXf* +-p-u$]K]νJWgM=)F~j_ɎTF鵋ZJv>Wj'vN/,9{"ovz􂵮׀fĚLFkXdǢY*)kրZRx>_,,3mGL*smɚTVk4FFݐ%$Қ t|H胹@x.g+FP+mrsuVhIV_2rZ^Mbf9<,h?^SGms'-V@5?~黣>7v$4'ӱ?v"_VB?*> 1/=Z^PNFWQo329VׁJʕJiNXgւȨ<ȘJ~SU:9JiMI!HX{Euq%Vm f*T<"{. z$G95P^ǭkER|=c6(;{#/9Re#/FVpHbvs1rXq)"$Q~KidG>z׺ƞLCR ҉1D[Iꭧ!3ΤMm.w 2,u}L.[![Ih lh]ARr/EEGEtRr/QvV-?,Ũԏ߅褫 1{ԩI-9h_vs ݮCwjRdjIֺ.&n9nؙ] .xs ^0oAи;q :o8DKd=`#x bϝ٫{ؕR0 C]K+~xkJ> r9Qy/)yCG2^d>$66Ԗk=rDo] 5M~eeMOU''WGML:h) X7igYn0j OeUR0+Jbi3Uuԓ!k.w߱*Ֆ|F LC[zT3vԢ>륜.﬊~XPX<O?Q2cVEL+'0yzngACƺv-CM&/O[ r_e\HY`aV)K gVXsZAVy.@S2e ob̅lQ٬f?p[=Vl:5rk#f6bk#f6b֮j?2<;p<ƣ*.Ǖ,/.oZ.+\|~tqy}?j"|my%SqKyj8u~x+3V:z &fdT^*WA5ױi\̪9'hR^"խ Uߎo2e "g?fHZpy&;mAonFoo/xĝݜ=]>z :+~/b3ːŌS8Qƒw {\on0ŏ]o)Q=v΀*r14׉yT.QЇpx!_g޻BEElRk'E*Z{ZDeJȆt (1Nxs-:G?ʊs79ŎUswJa)aRmH-@H+j>A#"7ʿ"`7K7 4Mg6@';dي#:>dt%dK-m[*㻯qߥ8QUlv¹H=[xvl6½_^+|<[c;C F"H ZNd oЫڌ]>7xvO+soM-A`s3W,8Uo9}dokDmk8B{,)98XX-W1\sV Υ\̯ 4ⳔiER7?=ld95w#𹤝 R}>#d$xǵ_h75@[o :}^j}>]hiPxFxBZDᤔ.y5Ce1\ozHC5cۅU#L$>caTYazh`Z/ZJvZ%HHePD/Ш 2]a 7\zށ H^bި R*ɣ<.s#pv ;c6K1[C[+8J.5e" ʠוJAitʰu )$3T(eguٳmv@.{4i,uY+\jhxr8JTh4EQA5N1]Z0UP : P5,%p6jBH1(7^jsؽ$ZC[+ v0\hPfsBaXa1^` Snm dJ6`̡ #:9T_;>ߏw~0j Ly+2 TmtjACB0ؼfR8|(Cŧdm,uvߧK<-$\HKF$EN <DɒE О=y\!jɓWYh.aE7nm?v$- Ӄ.a=*TXzT8=$}icFk` z-Lа`ū%k{hZIEof/chU8t,#';cxq `dmڒP,Sh"{|v߈F֬]KBҫdgʳF9_05Bqزᮭ\hd+y7b YxH7:gQ<:P8x _wFy, ^Wߏd$'5Ώh2d {#_k968'#h/aI-hV55T| 2re%~[^ :-:uO篚_?Q`=nq;?WwkvS~xe̡R"_aDS#MKJW:m~h *;1(gFv~EUCL(] NE[P܆P!BȜ,1oY +l>k-0IEXS%p7$f_dE_!N+(2<:q,&dzVP &[CUs$v%Mڴ [15VR@3:HϮgF͓r-9w ^rL30FXage3%)+|yL1gLJJ[* J 5i)o-/,ABh3\ ]2],E똉m A.IU\3s4D6d2ϴ9%k ՆPJH͔mR!.H1餗&$IB,ݸ?"R" K>]^Yxxw']3xjAQؽKbޝ/^D ZSWxHHeQWlj,`̜KK-xG,~[ݡvK[X)i ,em>m㢺*KzdY0DШ}Юu Xznohr+9F:(u^6yS5o)&;~ZoQkh&j40Y];PoPP*E>⺅Z͍ JqI@z%t(,Br2f}ݗ6_niW:FcI$J D͌+Kfyt~JRXE6]A2cP-9FMBE 7yY$#5EFt(."=/TB^$T8IT"M#Wivr)#X֞Xג5]sJpffB{B52M*}=}>"ۇzNV\{T߽7>1{^c6–M{z!YuL\7y;8>HiSꝓB~>ax ǵ_c75p[o8} J({(ךI\Ql@7L 0v&. HHNۅxؑT&6FJ ~h0֬l*oZ/=AXKs IS.yu>PA: 2϶Nu yҋߝg{:W!,iC';؈ !i1(|dbHK]U]:b 'i| -c)2s+&L'vf,a@9 yvڼZC#4w.b_rZ*S ]fQ8("$ C#C5EQ wAiNXx~`ak{TYAb8n4"T;=Tِ.y,hCYE~9 +=6Br9g*VPs!6t>ylhV/_EiJWwmCǗE "0T0b`N[RDw.{DgӚYoqɭb.3PS= p 0^th݋ mlZ?>H``Kn~و51lABE?5jK(Eb$c3Q0 lGJKV6ǙdkLG@\y?MJ0img9y]U{:9?hIoN~"'ϗ `y p7UJ#(F.cWv0$- _psRHEMX]'n?>W׮{젟{OgUUUt*MlYH!?Te)GN)nC2i ĬLઙח4k;{{OT`T`F2;uqҠ6ku[\D[#0N{N{Q߶Bēoz_ٟ?=xFyc+1`{w?x_^/zX}8NN{on\}zzfe jH?.b_$+l>3GK%a)rT M*x#. RKm T֍e]bBKLU;~nn2&U@`NN\~8XE!t&!r#!Q*kXc*:6>IUEǝ@B]EǝhqsHtm8z & %']*IDs1uZq OPEaltދMjm*}%v!Z8n=Lb)FV5)lE@Wj-/׆ʢSuب<:t$ 4Ti V1wrF:": ZfյZ\IB4BZQ Z{S$`/5, QHMTRގQZ >ZPS^Xyu /˹`x" "!I@q/}4/&k̉@G4MN#+SQ/gzFD5: 3w/.8)W>)ww'_NV΂iqJy4L9v0|%S3ߧʼ=STykVq 3Zɨ|R-I]egm$#S.UQ{qe}N>JhTCupG;?rh*/oT 6)Y>QD p=>g^ث3YtŦϚQZܞԶxΌ&R4+xk)9xӔ܂$D%l"2PSr:2IAkKTl' Dq29ksDW!1wT petzv{O^]k,H0+>Zfլ$D/)){75q"ґ/Nc" 64Phn&0gRJQ--"YtH[-Nsцw\F؏ףè#9ЧQsmEo 0F/o;#ĻA2&ۘ8 (DF9I,QMK\sHI)S0h8RA Fp/?:kV 0Oyik-\weox+|q]sI cm*klD!Ͱ`AMAޱ^q Jk=e6\beĎ!W f[TVap-| 9nN63d|\WMZ hSv*6nLwϟo - $mT+0+$s{Q8s?D׃$|@]w;]wNZMJ㲻'^\ū,xY3_yW`>E(;\r #sTs}Ipaԃyk`M m2͇I5ЌpڡpT#ss?8QʵRSfDV>{evRZz2f_ ]wBiLn,Y6`|G̰ZnUXhI)IEQJ02֭8q.E4>Q9]Tk\vQ9{юhzja[H!5m4B1~\y^Ԍ3_hf4.r9 䐉[xː$Źasd8J0*еQ00ȏ\3,)抛-'5~Fי5꼷VbPA"Eŧ҉{p8m(m`mR0i/_:$E]qE>=>@`ߎ5xцB}?s3͋tA 0ešħm9_,Ӝu]NEQ2Hu_ ]_FRFud/`>;z{{3|?U-B{1a; wa:Ubofc모7)){'xNoDd)A ֽdrq=ђR y `g) {a c6\x*9sᙳ˓oNP7;R*wA +Lc:&< *, ń[T9PQv4cƟ~2@v0{!؛~}4,6$#VTk$hc!*pt+ReA6~%WN[;,Q57Jx"|pYTSXN(F' ᭢`v*q3hH\6hq T%&D587Reb"rxሸ"i7!J`I71J(:p1TTLԎ}mKZ"#KLi0Dt"ջh&:@5N[έgH)>TA0w#ֆXۅ@Ez6IiB`qpO *ISʥL",#3)< ,l/J-_ΣH<1F td\+PcI\27)6V4\㣗Pk|:Tr]dT|9r1~u}?5k1H}ZU'繂kFѤjW>d*;Dk1jKÄ X0尿Ȁu %[l3~Ciatk,׌v؏(Zُ;Ɂ&]mӂv!9v:lT|댄Wy@Fܨl C& gN*+5v`htlv_oøjH\*v:[ދ6E &6TJa%!91I0U2b;O1h tŦhNFe$&s?;h>zX@)RuXE)RJ$t놴ӹE3AՌv!n Vkh'9`[CLCڙlvn5dڭ!j $t~N;NTEE54;^rr3kY-G`״EލdpG09G/+C>yq73]ؘɔ5m YN~QC QWW+)Rqore,ע \ Y:VN[H`v,AUcIdBx1hCNUt"u,Z$9I,M u vv]zz*3oTUeZ!mAv|0VVoWٚ!\fz-ijlO wQ[蕆?gc)6W^O)7J^2L/K(a܉olX-r%p@m굅f#0ՂNjC|}҉\Yk: #q d52*=Z~A2X1j8.3a>yհZSx6mB :j?=iiwBZ&u 2AM yD*j4t;\,6sV:^a[0n:gSVdu~ێYC% #P|躟E۩i2 th׻5tnn YA@v܊sA~=WN+) u3w7xTѸrBG">} A- n\'r}v]{tu;?k5U[7\f7h!ȊKz?'7DJ2৒6)3 fh6 X6{?Gz:x]Awdp7Q_8qrdaxI&{\t<ރVC!M7wB7hga݆6v񽇌.(OpkyZ!^z@IM}mɣq>~-d=x*Օf3cYdf5zx0Lwl a4y&?ĵ&oT4nȝy)v>^J#WqȐ[4iB bWn+ &T y\h!1rB 7 Ad'gDVI9WK;7U8:}Oޡ#S%HDN>Dך$v`#N)#vǩFlZ؊+4#o͔72rw>!;O݇.ևƤ+;d^+[aĕL{yh{a"h04 ?4'{UO\z \^23A;Q͢gy;O07bs#t ryV$|lda㛳^r^yƯiU*}*M^}k+oG[[2O­I> G՞;)ͤS _wrշain Cm(6!h$1߷wCW7-Y~Gw{M}1J_6C"D,VƮ74 }vW;B,Fmnn&pQ袄qewz_ F8|GcVrqᡝ3oa/lJw ]1l ~&86֙ӧwŧ{)d ]ϳm)~v"ڠ,LȪ!U WF4(5Lؤ : TNEQG(TO@)FSnM8J-|TN>=j&NLUEL ]T"5AQ$>䌩$WdEĠDwvrRȂg&;Ϡ yj2soLJѡ%)IVpDE EtYUbⲕinh0 7;<|"m6u'녀'C?pXݻ]fnu,WEΊ:z,2=fߋ?K yXe$2R1!/+˨j|\W4&=ž zۥ80 ֬:t=XhӔӎU5=cg7cSFv :g CW3:nU3jz k=8T98;m+8cmG,.-"$T4QoTJ\jc;fm㘒i͘{^#G RsntLtvy薷^.4*U\Wc$8Z; 椬:*\6$cE/ًF] Y2*npsOg m[1^fL> oɹKUxN, 8j+m47$v4*8=EL@p(ɓRuJrŝvDYYjWl8 _9r.kZ*9oG++1 ~K@RH*$ BงRρ7Yd;7嬷]vT朤ЗUɬi ,U=}4$)=3X}їٹ+p1!!"4iy C\ftTlZ0xZje/IO鸻=>HOz{"׾:Fb67|_fŻ".'_]U riDy?_uvi#-a{QEX!\pW?.`B sa Pu;8؂AR_bh˾[D~TQ ^wӭcX:ܺnt6gϊ79݆xP(;&Yה/ 퉷(mk@{iKfL#ۼ^cdI2&bW zTPu:^?VnD_T}] Z+Aw%hn>h1{A]T XJ/)KIq$ ޚ$xB-xNdj,ybj4"M/OzbvݨG#_z" rr1ZۯI052鏻Qn&'? 9'gS^ .N.E=i2!x>Mn_?[+pDk2rfwo(=tlƢ|WgwCbŒ)']rek6yi;oi@6UR:qjӺ03mDYDb*jdż@1#*gVK]Icv-9wsRg D0d)yF[$!@^:ML Zʞb\]]_N)>i%Ijݺrh#BmilhQ)ATa$s1~)2Ht3\GX33 L:Eg,y8`^t$F`B/&Э`"dy7(aLVV) *ÍI.IQ`ri\lv:SH1ӖdfCNѪnJ}Q #q9nGwDvyE3[VFM['Fl#F;F ŨuOnlS1$U7bԨSn؅7ur ra~\k *MXI-ȠDI >v-8z}pdl89:,4QyB 9ӡ6k_ZıI z#-ѶBL"!c$Ǭ״5Qhr>Q%7|KmYIT$dL{[JɻYurud;uVOvӃx7-/VNts.LO'^W?q""^R˟/ꖞ"S"J7 /?4c;zy5:[^I3od,̪*822"2G$-ʀ%eŝ*>Z=FO˔R~[/~URDQPaP=3_s.eq?pf鮹`\mi\Lv-3 r]uŗwy9' pfQ` DynqЮw\Կ ѱWc7az";"Z".)Agй#YE\3S#ԂT1"*T29b P{c)Y QµGzuH< sCDy zQ<4yI,KA_ObLZRc'L}=:W;ڝv8g0|M(|4ҡQD2K3f^X0F8甦,ڄ! uqGLZ- ҤV\{(cǫ*Y2`kTu 4!08Byjx_!#`^iWN.Ca!%6ɍFI$Fz2;&IC5$=\k> `u;):1໢)"P`3["y3Fȑۦ9u.jO̒|g51OO`js~J6W'd\-TdTTdL10! saj$(JBm >: QCL 1x^FܶA _eNZq]{"PaZ"`.KU D&謡 r~yes V%U \7ְ }TtIN2vJ@" yJxrn-aFdqqW)P zHP.3zƶosBSApWԨ \MJ %a␯6 O|XsU\y> M&=Ϯ4Ј> lϳ@-ZI 34) 5{@O lFnƧpl??7a".][E/KwVtkKI@ QIG@I=r &>]V7 &nor]oUoLj{0p77ꏵr' k CU]d%]dZI B~cfC,b(̀ 3*(<:Y$2k`%no'{M(|:k#*H<.sW:`7094-aFy\=v~ޗGQd:M_H}`Bt!0C"Sά]4Of ܷeV>ZVGNJQ^tEZ.5u@'SMWa4Ƙ `bM`DZ)J=9W^ |q (NP$%kNcmH4nNGm!?E1F6hвcŠᵄN(=ܸ?Íчpc;!jmq1(^"0_[_a~UZsWG,L8 XAEjؚ^hr%2JdAWYU$#,&TvPE,Jϵ\SBb$&I\%(/W *05#e>sn >Ih׾-Y?@Pgט'0K8'w~1Q3YƷ a)4dVԫ` ݀*uIGv^j689:&$@mb7JKb]w' Sn)GL#3#Pޗ^DLjShJ~ 䫏ґfsJQgR$墁'xjJ3XXJT)@1S(vJ,rKݛo}d NT VsiRT4rfV1÷]+Nڄ]ڄL lhZ*M{̆e!8+gQؽ>դ& J7Wذ]ۄO@ӒҷtsV !_9-,Ci]gBq]t[hk tCrSJ2,S^w [$ҼY6+e wbr@d`.awpP{>בּ?qh=L:g B!u{oh|YSWNOABpsFBu9nDv}v_iQ) Ҁ0e߈nRkcEr?Tc|x'\šF$-yZ3 6v1 a[9Wx;(Y8tᬻxVDlYSRIv^37BYG[+[`뭩AH+ +$rg ^t|o^5 'rrEG[4y#0.%O`DTyB|9C`Z2? t5 2 &d`\ن'ťsUwU ]v0% }Ү!<1K֕aQO[6&Ha=>x9{z5o(sLٔIJG0y 6F˸?$k/qC6nS?,ٶ[?ZbYԀܭ -ZN[?G}ѧ}ѧޗv b,(M"AЃ!Ƀh<~1S/^C}#̪,KjrKM֑Eddddk#/ iZ"LGODۚ'Mg^w+_0)}:Cl/|ZuH_]ͧś ̈́O˵ewIFSK}e 6LƗ)lܔts)ox{/*hFn1RK$ƒȌ"%EE^Βg$?_|SC_~gy9u~uirœ;>},HД#!FYfYiԂ2)8x^BLgmt$TIF1I8#rý8`*{0{t%2-zLݲP.?x@QŁ֛aUØD<}6+ _WD%HƛrҊW0!Tp`Aha=Ȓ@y0D$#a`B;1 ˱%B^1lo"j eU[7jxP%-e&jĬKRQ\Zߐ:h^!,FZ[@ +f KmP4C͍ƺGA*oKI#5V 2]`>YtT"-^ j0o] l_,,%|6*Wk W%7u|Q eU[=Z}a3y7b034:YZ"*ܸ:Fanouwb[}lZ^xa)X$2(ycqL0Jҧ;u`n&ފί? ,'> +[$䙋hL1G+n1hR rDmvDZ[1ڭ y"Z$S61c8L:VA蔾v;.BaEnńj>$䙋hL;S„4hYHq;F9"Bm((PQ+6+sSi嵸\O̗U$xuLwɡprex$ *T X)u%PT(N'?fًOfjj_^"9B^<%cpC|^i姈yB,$>/)s71C|V8Γ'?f#1ʆok8ž2ij.SniySeO|vٓՃsq0*{҂TjG6 5 cUkh?iS@1h&I^j`նevPu飂st5{H5z=䊵 o)ݛ+P8VCׂʳ;T "x=Iwr_4j| һ;T T6Z`H+pT$6-% I./KZRg2FRJ CcXhtS$$gDՎZPHȾLSRQ5iI+dwDク$"q ,]"$Lv\H :v 3%JK )Ij.G\v4H1B<`cLvR /ҖXEYᤲXT2'qۏdZn c(i l3`qSZY&[> XJq =. {yFoEl0@w :x,@|  ՇT~jS7q{܇ V"X{nb'RPSL:Vdp̨1v*© cftP*#* {@|pO 3Ԭ۪Xt5Wvg K5,!Ufi'`f|ay5FyB4 d",b59\y}:D3yoC** 9TVPe7j|0K"AGO(h' =y* zp)º[+(5$,E R<a2:= &:Io[^6F2찷 "4ߙ LکI0xPEF 4p-G~߆w7׵2 <:]Ƞӓh8݋殌>y LDw&u=ӫ$h2uGZ J EC}'C=h-0&*ͨX/ jBJzo|x܈`y|0$4bnhtq7o*f_<_!w-C$cFd1qq$]`lh ǻݧ| :GX65,4||TzS?^ɉIN'Iә%KrrFi%4R(@` 56+`GED[1^F4bsS?)|ɓ,%CJ藉V KRf҂ynT@.B)\ &ɍ=1`Je8a$(FQ1sɺ,-()YuŸog1_,$yKr}}QJs-T0@c@.V 4@( Je i%GsP.hD~aN1ªd?H%+ ?ae7䁾@Zz4ԅ_+q/eI.]&)lJ67V+O:shR)h0"(c̄':g'?@/&oRÀz]LV5bdiՂ`>lo>0+A8tt#Gd5QqN0G0Vy-sYLRs k}Bdvp7|뀣TJ6=vE[bA%Yl B1Rm4v\jn GDVf.R|lvs+.n\X4gU ۊM)>noD=V;k5SVيH$:^~Þ'INxБSk[űEvxgr,,K'ugFϞ8] Jc,Vt%,V쑳}׈41}woAOtOKˬ?g q~~L )C תѐڟ?JFAvz;4|*%)[-/RKjF=O`rmp=[Y"|9QFSAvH0+#jܕK" DB|緭[ESWo`r8E!Ż^ԫb^p]ks> ̃sނ9/59fsԔ9ߥtU.}G'S%k]nd6CBv)ZS%"/XU؅; yo%Lu[ͻʅF5:?f.^Mn9=]Z=\3ܹ7:}ʲ󆬲.ޗe^SvG욻e~1 <tVĿhC01C.1N QmP8wQVIn|d egD)׉o,=x=h6V?wXωS#n]-L܎{|S}[sZMۚwymü{FsM!^w,/Vu4VZLoFYtn< 4fF`֗_-_T.{uposSV*gW^.]48ϖt0I$pcAwׄ7x)TG^]}RH;QB % @DǞ''H#tƃL~Cՙ&5ٻq$WX5' /gj.͗rXKdj5(٦d)Ebh&%Q$4hҎsH7= jivy- Ю8Fj[jwc3Tm->d$| HeK.q'H[;mTlnЅI<3<@}(҉!nTUUSd{8c0)aS4k aO6.Tㆠ6`hqmol+X&d?3Mc;mT\M}ً 7[gaQYŇ^MlI:II(^Kh,$=xΓ!t̀mmьx46/KrtdpC@1JhdulCX3mw+iko:6Ed`UG!զ$mxv[7h䯢5[/eؾj$5,v(iJi:ΝВS FfTҔgdyeO^ ᮦڇ }1D =lO<v>C:p趧u rdQy>Ƒŗ_X 3cl4%0qLF@?P1gQ~ԡFs,?!6C.2S \ъcQXL+Eء[y?^9,$PrUwXUÇ7u${XņJVZ44RQD8Tw"]jҕ~gs"4=)aA0:YaznE;p@wʠhNUm\&6|lXY #M4ej,;]|;\8*UWNdٓ)q,b+T՛]{2^_%J #z.qO/C~.o]xvyS t>VE~F TLkc,Eq2x2YsKDeꐒ^$ڗB, nQ|&utl2vJJzۦD?w meqC=eq߯ϕq*Ee jDpr JTT8Jl:ٜ CL]roL2 *)劢*).]0vD$ d$a`6BHeFhbcC2%)@h\nUve) 0:=è똴4r)7s\F GRT#q&֤cWK\:\m|}F~iѦfW.V(A%.,];>57!/a>t,e/s0k͕Bkreō4jDcD[&3S7l?p\$]?tyF?>R +)X?};tBM9{>*i-)HpU$0w|ikh@j3; )R/ݬ85'̥~}=_{3\xtu{5Ofly$(dQ5g5CP_xE9$FţB)ׅj2Q|uP*>natܿ?,*YXyG9@ 1*3+ݞ@sShKӢXK1xܣ4TTڎUx~9Ԟֲ+Dd$0:gJNI*2BBG}sڡX=CEDUĤDWVx Y?[QÚ8YsZbҔN+(YHSvYΎg=OѬg^㮄' (`A*;X2:Wc+䝁9*'KNj>j`u~.S$=˴J%X 5qVѠ)P $f1j#px1] Zj!e5C(bYJHH{˧[o3"Q,WA3%uKDZ/:}$ANS*ÜI7V1NhYp2Xj:v%/YhEAC&hU< :=Df~K=Ny^ sHɡ+9᜴J yz +TGedx163UlJ ֘^z**Qv8~v~am*lX-u~>lِ`Y/(6vUtJyy5ՁEul0pbNdecߨyӫ¬iPX\%=_ཧĺf0g)(RM:et߯5WQ{.-$LFiiw3m|"0غOܪ嚝%Ry]>aaASqO/@)UMqV[%Eu.’Xm2ӒK96/+`UO{X \Rcx} D)Q: RVQˎT %a L5ªU/TGN^;%Rfި\uͱĜlh :r n2+6u1\qv*KRS2M n%%wGi *rhiNl/.l}{тX\_ =;}d~J5٨Yl+n_-WTL$>  UBc)\M0+ŻV%zpK/[$p'GI1Wf"ŬhaR̪{qH1[/>)f8`|'Xv'XC7ŭ1,%u53$Ў [j0Ba<4lD j,Jŧ2:jUUgA|"Q%pv>SΎ6>'#>C!>?"Il VܱAR0mX֕DJu~z!FW +XpQ=REMͺ;-I!̥a^ͅ?C\|~2fͣ'>0~yv ǷRR^0ƞ'Zpon<%Ju%x/s;^>7o7ޔoʦ\H%R)Dj0'\f:5J*I.9u^&ڳ-[iI]&$KmbVnejk?(Ke$WJCq&7Re2KaƝ9\f -CJ`SY:)_6AYVk*¡(wNב~^wag:eqSe&rf&5ʱŠe:YZ¡'Qiq-K]QQ>}F`AraHyŽ[&"i ) FtkCYf\nt(`9 Kuf0 9\i}O:݈r7 +wgYݧ1Dޱ}Ƈ\؇- \WX _Vkro~J?-F1Ft\f?cw8W#Y\{ |z}mp4{v;ŅR VcT ?m=K 9aB#HNsN'Q1i? l?7‘扶pgp 挲 kͱ01 \ IUr*aI)$V9b^W$kܽ `FuҊ49jfmR+5*w麞Lހ[%w'ZygK\mZwO%'Jj7{ LX!1f*Pi[vCxi%&}X15Eߞ+ _~~V!V!kZa0cc ˢ .ƝD n$u)R J 㰮V38&ݝ]̧w!ÄÂ#2blx6al8TNFe*A댉 ,"GHTTnýݗ7q()O#gcx\2sd0 Zڤ>Ja Ծ¨?}m-,X7@Aqg1GIA$'TRh`THfy*G&\! &ՃיS[!dʤ!80 Ta%747xes8\pYzѯŧwJRVS0u^굞ZӮ'WKRXSTilKzXԔGNFbg;_kI=-5'c0h0H0l䕸>V*Yo'$A"R?d~ &PkגBuFz5b#KuȺc akWi^{kz5㨝Qt5s-o&g@9!݇ei䳵߁,n']'yvHc{X& *g]p}\,\S|`찦c&e#E-B(&i;?~;_L[>洿{)Q-~a; V2U\gL3/!t9' u\i3郝6=UѤ CDcycOmD!Z-b!iG-ߨ}([UnZF=j#X01 f>!Q#\lFdX4# X-Q B6ZEQ|Xgw!IOӂ!KS7gɮt+%m|= i'ȓ>]jXt<[]@.So] ƨ`AY.R㖓"0/NY2 <]bsftsa9ScK$<ŶԬ#?oOWo]:* =I앳 ׃xe^:]RB4iךֹ1"_zMo:3=d7/ěczaE[TW`2lTت)ix1 0:""~ܳH◒[~ؓJVE 48f ?&+fG"&kO<_,@ 6p #~k2Lzvt&t/@/@($‰oVF4<]l1W!5l=@u")FNzD)؀YhF>*ҟ_L3r vb1@?Lrsr` d ʥ | /tp\v3EBuwe?Xe"pk8h%^tD~en泗aN))gL.ϓٕCv`<|2-CC piY=gy^v׮MWg|S:Z\:>l AVM5'!iz6 -(s̗7L0NF_7-atPRk>ͳ6=z<Ţa,pI]> hɐ+,/)Br~4h(rvfp%l#94Nm5 KAփ^_m5\_ws]rp@ .Iv`M5K`jghn&.pE'l4U!+CPYJm$sbQcg&#OBg4}7PmTi@ٽR_;;fxVFEj1̚iofަ*9W2S\Er2o;tjt MMۗ_bFpg)jqVI>iho#,d|+BÇ@^]B5Mwg+iMW!b*fe8mlr:L.ImM!VҿKB%Di. p%ZдiWlZ pps!TLw00Xak/(}hwY֙\Қ]:I"=ލȹAbbhSy}M{vqƊ3*\lNi$%FOϻg?9u5Ď,5^wt'4Hڏ An9)";5Aw:tyJ s C4 #Ti[Lj7.UkYӵ%j_/ӹV$ $׊Buuzuը<:##N#]uZªF-65x?DI0L?Z"kkBk;Xpl4$X bK䑣"A Bh̶bĕ=D"T[DP,Wr ?Ȁe6ò9JNFޤ[{2%ڱ@,Mll[к`c (+ /*ʩR̗Ԅm]^G1Fbja%̴R$ fEqȀ쟸z;pQ(ɪZn:G' V9vudGLZve%kݍkL$4AI#yǎF8L&{ n!8)'[V'U썗+_B whr<~<[ޅSp&2;?/|uqcUjbW MS0 uey禋)VEm IWH A's8U9 /4U\,)_֠ɭWPvg>$;s-S.4śbuʶQD]/hĺ[nUpg΢ex 8XWne1:e(ݮ4%h^V|,ZzBG ڭ,yXlE5׷dխJ4vC>s}`LUR("$<1}~Ucsi~12Pr FE}X$-<"-t?F~iOc]51I0wO5:!KA \Dzd8ra:]AkXOop?-Cզ7C?r!vy~kO6n2}C+wLnt kk?&j0u> i@6 ):O!%zoodE1L/Fބv֙gC>N@Y&{=[5&}^7h}.Q+=npNƻOw;s@6^< 'Ia7$jD C"~/iY}D9]>hQa[!E*R.*BýSs(u&p۠'it>rogh]MWNc> vI31V?}ˋX.{0΃.~i'/Jrn :ؽ.͂% dˎpx2^gZ+f}>o蜼+Pm?- =K 61(BbL2<9r㛴HbZJCL:㐊&=΍D%rBA_R7v_89YPrj̝''0u m!XL}D!ց/&~vᣅ0QI!sCK2Scx &J[ 9y [b$Cz) qV#ߓ(;M4e`Īk],LVG9/pL>ˑƷT"fmkxEDB_ǟX &i};%K_ԟߌ6kIf~Wə0HFWb5YvEPbŔJa@#Ø>Ʊ% +ᱲ6cbc-E/aŪbT_k6,e#v,&~Abd y) m8`[7 nF!"l+6~':N㱱vNdPr00jTseLigIw2] V5@gJR9 s$I$GJN>AFػ(6HFT+f =cE \3aN(> 2 f$|1 ,)xAց LppSt%%\r)i-Imc/w9}qRvN,8-601}bh CrA#ORL݈2$"Ս Ly ` "Pxc&­yp!jNèk7۷ֵO(2ݴ_V.0vXapC ;Թn"uWȎZڽ3ұ21Lv:i1 8NcI(} 궼)yϒ ZIsK8!<=-6)٘gS]W~Uj %LhnT!U JnAAgDSͶ,Ӣ| $Ծcm ex8l!\@I!Imtt70D'xRdCTQr[dz9@RX+i9FP2-4[mt2jZ}QIw8-3!*T-m%SΨ'pFW FJʹ!_wik cCO_['Qv;7:W)-WEL*Ý=; #bW\Z0zh^ŷ.9ٴ$oKod~ /U4j A7 …8B_'x׳;1 3Ah|=+&"_'v/xF`]/dɊ>ͼA#p S~[|'x\<8v'B7 ɃbାFz܆&y8%Z`KAb ֥8ACظyH=NQgߑBiĉ@ @ε,c\|k d>V&/8wÆP?8zbpʐaNeߖ%;-sAH4`|!3A ,#%4cL77Jw2F8(ޅ5Z-@5@G.4D{tmH59C Y!HbN%/V8k%f̧W_It˪u_ٲln轠^aڵeW^}tS[sJ7#{չ՜r[xB6X#[{vkּ~?|ţZʑz̬֓&罱BgS!BnMTwϧݛA:k،1PqT>q=$ *h쵷~-\PAkVd_ii(W5-7]uQ >Fp)jP9E{Ɇ禳㐒#QNhڛD"D"H 8GDqx hy>-\L(U޳"u9Hi C gªZV7׵z{Mߑ1(XKrM~Bxo_%T|Vղ9t ?8B9JmVTXV#q%Io"iZTOΪv;)]W{♕3㏿nR6&(w2؊me񛲭Ժ@,82;7Fh2[zJ} 7wqG;z[zizPN{xDI9DTkU|| H꛺ J D>r6t11?0,HA;C*Dըnٺm/i~eHR@VFou_> ?? }:Oq*ӎSmnDžJn/׫Z*im=Z 4KjC;^0BS)ٞs즡TZ^К 0߉$Q-feY`yeu3'0 $axD|$`J)ja5lϺ- gKT[MePh-o4DTWGm wG|^@N`õiL$"TJ;[`H\Yg3 $3˔Ln볤";'/bҶiLP&߬q3mB{vJc;ݭ3:q$IQtOnŘKB/$>X:DB4y~I5"](Kk.3"U:Yx: ,%)D"@9@ gv0t 4_1xv5-)6yw=? Ԏv:+!$BaK7 'LB@7j FbOn@@PۼԨѺ˚2,I9 !IZ5i@ͳP]:yr}l;$tPKN20u Awh}B 68uknQLS{HOS- k.xKˈ8Y<-2JoGJ?`iCj/˃PC;q upXvG@D 3O.u2ǓIZmpth\ZIA( υ{ #x*k~ Q/Ne0K=Fc(k 펣 @x- =-'P*"A+K61n@fy֘ӹ\ ) f q8ǀ IQn|1R7.7,VOXĝCZ('Z5X!vf~NqS,ς?/tHbNDa*+Gb'|mnxA^4' ;C@پ^Jv[Y[Z΢/kx;ACO`A#`]xv ! viRa ϷV<+R7O~+Gr@Zl)!1PV 3X= 7 IO#@u~d[aovUwuڡz{"€?Ȱ,@1|M?(Vv77#*CxNdAY@A}[{Z9 $rٛ}%Er(+N୛DD(D'9oj};%M3Gm`aH-ys1$ϙ0sYuH*0jͬsu9:}y ( 486f&M5Ɏ`$&e!1U,5{A݀. 3{ JIrfͦ; $8#mE`F "*f?.&\RLXҒ ʅ&Y6t?܀|k9ACnސ7a a\Pq8}K c݁hXFctr(Ɓuc-ԦHc,  $R,?,pP)|l&Uj D0·q%!W[Uz5  G!%2LRѤbcog_(|{Z̆*,gn8(}C[xg[)&21onl 9]~`E˻&63Z71ߌR,[śEQ#\G EC.rC&3{—&m0e |!H!O ˤ\_Hcf60hE娌,Y>&u5V)}$WTzxsfT?GE~?Z8> >S\w$;eԡ|ҷp"zPRA:(u1?"ͲDzDYqt Z \ȾKg"(FǨf=A B +4Kx 1Ky/j Αd3L" dg7:AC|bU5bV[:F}'2N$7 z<-]RSk BHnr>ӄr|HT\Z(Y+cQoss! FҀ*U}I][hCj~ "Zp6knsmmEvwVYUկ/kAHID9( L&#mT-<]x8;]QWkڲj>Ã9fCг6NsɁS^*LEx%j3_*Aoہ8zeR=[zd AAs<R *GDlvtVbuqJ'9=FyCǦLRSAVMv0 2 *"[mIo*M.=j@tG"}ŶHSA(hKQy<ވ!thb"绚^ʭ;*x5*C( ~2P]|LLkv]awMs5o;=5I?aPfIæak[_cߥObw!Ƙ~r5 ShW3 xsw?z$Ɣy$5nsk_Qb\]\=Њhuumy#Hjnj`oz2nEMϫ'IB^=S#=5xNW |zfQc 0pGRpڷwMI %5Wq,P1oA2L91Q(v<"A t猛b @K,qa3>~ 'lCI|=r7Ԥ%싪<)ܸS\ 1梐{*~4)f/5V& gǨxl Iy?Wm @Z2u+UskW]I^6IKlt"hzp^|mǯ~H} o_Zd2z|86t7 sbbM$_r.}'K9l*XbOp-1Vy jyT,ݲtfΤos@‚Z40P\.N9t0$9 -C3Eonn#w .$+9/\KV>u7xs{q(@,~~"5 `6PIRuRdJeY)BQj%1&4 jRP9IeʽL*9R4mtD olJ"[wKKaFUj?  ?o.hMЯ8m1lۙŰny 7&wW 2@? msgr>,W/4_)%jҘyW/`:7 ,q_^] p6_,hj|WNgK5aV |`|w%S,߹w2$T Հ`DUE9{4CK-r/ypxc甎{ qZSPlS7G">p47swby}9@?!U}Y5MT\b|DŽ cI+t[~gylnaxx0RszJk[Z;XkaJ#{;+I?xbA%BtG}l- %D zuq7vPi`001.żFӻ?K_5.|9nQfqsq̮ë:v4K1O'eTi&3>`I Fsa;XIZtfy= Ckz+' .w%7sEƅ;&޾ͦi _V`dt0`œsAXo ŰbK7!RTmRUn(X;//h3%]R(1Dg(g3F>sifrڦ;689\NT6@F3^ɍ\IѪxqΞZz/%A1)9`6fE&`5X C6kV!w3nhK8w~` Q<t^n`3糿(o9+{giW蟜oÿ7 o8 Vl6LaƀUp:6\p9S04‚m"d|J СlvU pR1 !#RAz7?\U_w%:[ػři$؊)~pYA-_4B-0zx77;^> p//F6#U,d4޷W# ";F]븼;ST2qF"dvc=7ѱ$bC"rFYٌ@)OC!:E#]Q:I%N:Nw}˦0>gSoz\Ko/[xOn>?d6tN'iߓ)>@5=i$ݜHhcY7錋CDA$\ ,wӼ {ˁ\ p)Kg;u>Qx|^s;,4_ I@$6148|IĸJ2ҧOhjn  JB׻96 e N 5 &=oQ`I!lv`?V<b37>.7p.@M_,0JjuuhqT%;4"{l dХ灒L}NǛHdA/&`oYh=`/>G%aДaK4J#,9brѢȾ\A& #Fkj㤀c+j;-9Ve+4-(?7);S_qT϶Ӻ*z~dƔg>T Y&`1]͸* ^ 㯗f ehK};xF!0HBhBrFfrΓ;̋~w3 ؎Ji{ի3[k! 3GѴ27rG`,EC2L]J,! *rϸ(y+[ 'ro[+2`fʵ)t&$` ,FХv]>C 1W#l*}m4&HΌT:( &iLV0F"%m 8 jA $Cm78A&mZFi$F%0eRǙ +`~h{;I[-Sw|9G$ D=`]PCG~xu`ā;5)j[\%nOHArj}B}(>{υ)GXJx2JRJ e<.IC3۞YD⸟l%?Q|Q+4i:-+m @-J*~zK7K:ʽmN OFV L-V'Np&SR Iy'c΂h׫CY$٥@HފC? }{2B;R/D p+1HM)k0>EIA(vCedm~ctD"`HA*.\*K؍;KǣqXiZ+ ^V7ʼneS*\+JjQcɔh\tRrH˗p⴮B"XE)Tn1Zʨࡿґ;IcGJw~ǩ挗%^F *৑aܥTV 1ZݾaӲ\.zHX<%ۖerDi910llIji Yh-΁BggYk1sWrk-2^R" l#M(2+H!J1s\lRt<VHw&rJ{>k6mݷ9T^r,#)\" lz1N# QH9£GFIA nkbDg;/(kI0 w$A{7aL½.kkFK$ؖ&0NIUH&T#!m4 8%36c,+l2˩΀l: ^ {RvM-=fSf!o\^0+9ixDzqx CRDJ9nMr{^g|>dqUs&6v!w:Xb%6<%*uVq<e)/tgjZҬ{Fn6BmSC xk1ZᢚṈ .bᎸ3-6ƦlFDᖪe빀#]шuF՞jblF&%Vb*lCQvl8q IXD{Ԇ*]TMQҘKI*1PBO!a>'X1&df?YV<%)폢!¹Gv8]ƗL"-28Zb=j QM8SHuԧCA}mL ԃuC>]}_9\vHEҚiq3‘—HUL5ࡌRKD$NIeXxH+z2L5VԃJ`թȳ>x4DRMŒwym,mݘS3'ґ7|J6 ́#8&@޻;\ qÛC{Xsf}ΑṮBy;J( 7G=|Ro9 (4Wq(7G =|wx.M&jX1iI+ǼMQCQ<ʣ9YBr5[hUGDAQR1ǺM;]o7D9AN(ry|~bPQPIQ(BQ+%/"(EG-9#}~!Lh-Ļk3U62dWd5a:>pC U9ZI] ~6kڦ?sVRh3bI%Wi3˧VPw*! OZ ;MFD1{HkzX.rXBNAd B2DDylU^MU˫nе HXF$q1hmzSrDBdU7 `oHeךW$ []OZ PM?c' h1Y!!S>P*| rIz 9wUX'cmmjF#F.c ]rBomsBmo‘pP1u)S6ڊ|qC]\'0ep q$7%x\i P'5TsD44q-7>n 8O{`D9$U#! ij'馛8Iجd> 2CH)0rPAkJP'Z@CE|b6z&JvIeDiYzP+-DpSrN8L|ylQP_,g\IMRG=v1}u7:<|WJAVATHnt807[C)$Wlz\~pP]wkw柎~sWcѫWtk7&Q%݊R|;g仵]V13~ .wyMf{F[!ȗR1͖[>}`tv*6}Lirc`ĝ/`8RH;)S#ѝSP7ǀozΡG)@_NŰEi|Nmx~}O!͎?4'b 4|, hIlDШF,9 U; UH-ڢnVk͘HEΤn]S(ZOgi>/vve>QaZdhzE=zQ[.=~@{6Z.^P0cSGXvamg?\صݠTr} 來k|QQSr"ð08~isK*[7i0;,έQ;MDdҎ"Ys56; ٱ`5j;ph^W ]5 =OR,ץ;:( 6p0o\7~~1k/}Gcap Ǐ;#,hS-z Tߚqvݱ)\6ES_JYlB_|l7zHJ;qFD{>2pRX3K,b󇫖]n]nˢY)`1eX)`P;3u?O#D:i3U̫ u8>{& `JnCҨtQV#QC9>24n ;)Qeʷ: X~zFpKyqٶ8J %ZrW̅\?Z';|L.l‹6By)b+8"L<͍D:&_E2 GϚgi/$om'}k8q4cp=E5睾ކmV XhYr9,oˏ (X&ˑ$IZyaD/VҸqYI㲒ty0=>$SP ku"Hb[$+3 Y[a"2#8)Yy߰y8x[5-nU%jt8l 4+?.?ߎc7fC~ ӂ*p=de\:t",3gtJ)oǜ"J&kP<،Յ=F5FVc$Ὁc-#aeQ;vLGAJs6:E̥!K6PWQѠ7naYnӅ #L⭝։wAhJ*%KSH$^~֛6qBf62ہWTZ$$iԬ#z2fQtǬw e g F@{4-K_~Ĩ$[jbޕ"Ց@ 5f ~#ցZp0B!²Z@ɧkcfqq#e~zgWX˘dfF!a['5 RrVPwm5CឮS=׉9+c84҉Eߪ{di0\9v*5! lVtt &+,0hMC~3FZʭDhܰm4_6hy_"NUVr6W"QD`|q 7ߦQ\bʊ-lEoCMwШG 'HM+|鑜j.v9VE%=&?bwщE+4UW=G'F.|, xW@[ f8LK&(bb2 s]Hr_iFh yKk񤔑G8تbRi0:g!s"\ x 9AY*A犚&+OrB% #yr2. C.ŇlX BV1r'5ޣHE}#Dq<.%gwKϗg5Xqbb7+}!]HH#1@훫Ch ;knzK$hHQYw ~p%C~LO?.NS\nq7}'%$2\;LleWnJrZź2]RSUAkh1h]##5Sv=]J']Uf@ue%V]^ڀ qـ q{}z2h~KZrg0g2vw)Z$N*ɣ $O/>R7eη K3[Gw-Wk9nmiFH'wKB0R(dp+\Y@*[E:4>Jk4[^Nt"83'4" 1ɋiBD b_9q Զ\з}N֑jjYEC ڵXCF OYuxjy8F{ AET=` ejOTRyn$i,GA cKh% ey52$G{6PGCHC:IFHYKe TQs&F |-tDL6)@"ythM@.HN.{*CUO`EK%k)hP7c^ז8eQYґ֭ *\A jb Z`n~R,sUh?R*wog(APB j$,i.\/7]vOfTQL骗jn  dOы !O ӻ?f;X?{qbd~wp/\q|9`OIj7NIJRyp WpWU]6_aO"=x;'5ffjm/s\߬gz\ܯWg90go_n8yIrå$D|B gyz 4`hked`C/M_\ÄܔrЊuf6O {aW_.*wq\1##_/fRӭTdRт͵.{5hΚM۝: K[k#~Ͻ; 1TBuqvUy;UuOmw`91؜[͗ 4x\iCD% %A*%SWGx}ma aK;X۷p'4)깂7'FhkGonє.h )ݦ FKd(S E|i51N%Zca$t @Sːw!zF|X! 7mʺ<+$=5u96M).® IQ۞.>&@6ivUQcTwUc&sT ˻Gv|]N8ˬkTBhp_Vw)%hLp9 3{3Yǿྖv^sM }dl~@` d/)έ1Qq'L%hi\ b_݋s|+>Y Q yLw/F RwIe[ Oki_(MAƼ HI=7(IFlΠ$ ` tgǝ|$'GkɲM礌v @PYu>M.JR˭)X`I!u2)6!Δ¸\ߝtSs)m6͎&[Ynmxf(=ȳ=Ll}NK'8`Hn#,!Z dž״92ڡas寴9zS$[ sQkGPDAgV I+Z0& 8Z$ &΍9[ZkAfo[۟q{-?KPh**Kr$>R B^ ^ ~wG׫ I񣄛aqUub7lO׭A#(7U%n icijl,C+jWe i?v]?9ET{ .N<*X,/' ^n6jk? m+G}bY)ȗ/gЂ~Sj=EZ&ʋ^#Ry})<(n=# Ɲ:툴(cL\$c MԜ Z4Z ct ?uo\.VF$$Ro ,!'ܮ.ZCe;tA%GpMu|1wh(S"d%Km3R~0w= gwL%hdxI+t@ݭԠ\= #(`ђ0$ERϺKAS1#&eRh*d)p8ZΞhEP'tT` ҧSJnpE#Z͙G[4%= t뷢v2|Z"xHrd ?#kE]Ҁ&;r~j4935=,"w4cA|,)O5b5(($W*0uyk0;i#ˎ}7:,3 l~<0\ت#c<KQ =Wfrmdo*'/f!._zޕA1Clhӵ5uG 2>ASzǜ2u`1/4bQ+g`8ZԶ@̶=|@}w 0)@qnWecno30"hNɗvξ48A{6ځn<=8_Ov 8J%()K+Ą 8n5RJyX+6aɚ S2Ta\lyU#>90YQ58 km W8 ќuFA_-M큶 '.]Iۻ!JT3n1w6U 2w%{=jKqx z7VHoV]Z~[.I΍ѹ&7w!QB2'_rme˵duɜ>]ޠcHGQ1 , | ~1 Q-(7 e}pT9\4Q/q‡wJst<83[u"%v]rv=t9ܖ=Dq_~q=V1'븟]: ?Yqs= 9ݕa?d^]gMRFKSnoB-v{ZT YZjLT gYMbmkO,87llMܵ  us Z$9x~rldٞ$}Vw튌M[vDfb;ȱ&(5\I(Bg#P:딊QZhz#ưQKSoGҎs-Hе5gFԜ(tb:Rg&s5?7_: Z[Z<zV6|d+)ybdۓ篕TZ)L X{]zumjoM~]]^)ùQ+߰[柗_f+5>j5+k4v^}hQZTK^\O~|bkiNw}dQCZ@}!)us{ձxzu@Cep1\#_Vu 0f +ߣY0Vu?=j_+jlsF_=jA!G rd?Un6W`F泉eAql٬Ѳ88¨=+D GT.hQ3 З:߼{S([GM׋6< ++mFE˝D~h d&Ay ݚRGK&A>uǖlgщ4X|_U"*+*~.YSޏOxz>KUHl0IQ\TE+u:Gۍg}{w{: Dޞeq9||IyvCYJEZv7Ivfmu_u[*ҽ1Wn*=FNk=p8W'v9컮u粮<}f= ?ӓrG9*&m uxֻ smt}mnh۶ m+mEp0oRj >{KqTX0WJv̥TwTϔ3 ,R: RbBpo|K[ҭӹE zD/[ Ӎ_ RUHAxy1-IwxB Toܻb9Yo$~H~ K1!n1Gs6RD"m('W)CJnȕ_-1*9ZP$gRrOu-7ɨ@"G:PRP?0ʠ"A4x (kAͣ$En$[Y7v6l):\wëjH}6,C7[ĮǫbGb< ~K]gxت^O5K gV ե)2O q"8cE8*⥓b̐WNO*7LBI PJY$9MՈIJ fJHwH bMF(£䌇X32SEUxCXм/#Qd,'5 " Qvc "4p #"JOgijZG'YmfuJLmqC&HY ޵K-PKH Ѩ6 jl;oT!Y<&# V(X &2.)c>#L-7Lz0q `(z=70(`1(,AM@`8BG4A*SK ⠣ Wt!j0Ftng>;;'9~I~y/ٹIï%{Ff'פtV;Lbs=44n5XN7o,?<8ϗa, E*m(x/k[l*[_x?.`]ܘ}50Jx̘Rä>d=TEJRU'fNcw9W$ۋ'ԥ14fǺB]*bkn:7HY(՚kd 3g%Csh?NcdK3[%- 15ZL*9ŨŽ|#E  t=4,4#ȐȱVA"vBBJ&1iSlk)i{v=D_鬸Ă q| 16hƷYXO>M"lzW~^tPW[0r#s;SM1SMyjYW~x5&0)n#/夨87 3Ow@7 L?~VG7Xn ,7E"M9Ko9x,Qj>ZY\MMP}@̏:,$lo(`Fz AHw.&Cs lH6a"IՉHIR=fh7I[q~ Jz*CutSGpzBEaݏ|keKo #|zzl2 \~mX|i0_S`M|NYZJVgq{9^kz/|H" ?'F=9m*ƾMnrC&,4DWpt˰sh?gM~V9ETRlcbnI`urvT@/ޢKւ4h=㼇#l-ֈ&LK^DpDvkT_!\<|Tw7s@;1|ޭFԵɗÎeAkTJtQ `V Q$I :8"n)Z{<^ &>@!/ #&D^K&N ֹ/z$-^EɼFi#,y5̇hAD(.-& `. 0pZ`"z}g&i{,􏃿ZPB;//aw} Wh[u)PBnT_\և~R)D 8)#8y<11[_h]%B\ f `UB-IԆE|" H(bO`1'7~LüzdoDݱ$ۏ)0|$#i+u=ָWds:Ivp@*¯ fv=dLWESpkӺA[i?ƫ׷C= ?kv %q9lxazs7ao7]d)fY~̺SV>Hmcb@WE~櫏¯;NrlXzY%.~'n1>,ܶt5SeDY Js=Qkt>}Oժ Z.MO*5jB1ɳ?kT>ՐO5Ԫ /WGZ:5Bw'8mr3zӄcTZ ]dc]ZZա /5N0f9#kuj5N(9#kjTq(ʑYUQZ@ QXZժ \E֨9#kj;|5Z#kYcδ9#kuj{"k`SpZ@{xgs.rfYW`Q޿W-rd""k\ ѿȚ@-rd""k#ݿȚZժ TE2GrdVM5Zժ E$i1;{]DdQD֤D(GrdVM /P#kjBwi1E]DdMQ;k`9#kjg= sd^MP 5#k9V&c]F4Zթ [p Dn:)h vyͅq8 ~Y~X452#\G/O?/?]!\O[9 Ra(jIF ?][sI+6z|&dzVyYQUy0}$neF22ʽ;B ! 5<,Ð ;8UAO_I=)HL2(GW`sx&`b19rýuym\ ?ܘIgO>LNdKJ*- 1.DkrEiș^1 qkܡ0&W"ޠ7I02I2o\J?,{x9k/,U/bW/$eRLr-<>!mTNPH.QId?}(?z0(JcqJ"O;2Te̾JًԑzyT=l\': ڙ@7ʠ1H=q0HPMp^zXg rgı91b\8›0Ή(rwj:.rg*#]d E6 CahtG@ގ91Hb3QPEK5YȬt4D,jZI@ W5{61oܪE a&]7WHeԐlAvUHUYtܞ+!Igoܩ"sX69\rp,\s/.a66nK1i6\ą'AXbb.zoNb/"\0>Wh5@UOgjlÛA|Nң%z-2H61u1~vu99RHJX 1J+psϑ2cL gih#5d%³Mӱ K'5HNv~o`I=|ˇ`^(BD#F2M0|ELZP-5 "uiZXckJ2l!T8g(Abk816/Y}1ǚ(,)h/7p=|*\*;ABjҼEǔNCj)GaSPJ+B?Xd:"jbi/| $\hlH- m!ZiX Ð#s@Ç oSi_4yBXBV*òz2ɃR4x~#8$I}  JN$-8A%X#0i3CeS.Pɢ?psTMUW) GƬ1Ak$01H0R0 '"c 1ĢU,jAQW#0}zaKٚDk K"S< gsZ&bTݷ ( @K4L&i' ݔ#; D4m%`>k ]-G)}(}  B}0{ՙMv1MFiYͩLMccHn?Kޞ~ީ\0)oK^i "rD?>xRm| wfRN+p`קnh[UzZ|;>Kkè}3P0S !c6-/txkZOjt1%:Ll_i 0kHs_аk㮠ٹb2&\?{Y|U60`W#BulJ3MAw (L!NѴ8K(FON0]k%: _\z3plr<{\q?x81Øۡ~dُRY\9Rpڹ`}n3d?f"Ko&2 b&I&*YOsw7cLŬ&\Pv1:̤ A DPt9Oxk30s]([evYH\uWeB4J|tȐIE19y3(;b ي5w.ieq0?o~?M%o8 pF]sR'W EG&*dަ%J X9fŸ vp-盞T 5K CZ^ \0oJ0JM06ݒr\1[Mn -Vb .u|p궱l(ǚDFU"K% pmxx2(<fw^S<xd*5KIc֥]QGQ"D`2VbQYTv!.wk5)'iCVIAA,:Jm\(b 2iICI<Nkv\탽 f#Q !ͅѩD;G+.PIʭO HVK h+ m,((#F)5y16Uh Bvլ^3k.RO(i,Fu]PNJZ,6㯡S2 Yz+:r–%tePdb  .gBLO(86!p.{b=/GĂ+BeZ"SJ2>ip1)XƂUYa*Լ%ܵxxȒh=Uu(r՛4iو|*^ߝD)Xqϛ\ bSsudc?0Kb>ƛΡklE"#ϜD*êG)յؕrh6Yr٪~\@O^$[&+>k Zʉ%A&bp\4d<QLCy徯JlR^׾U Z?hC" ^SYUۆyuEf:O=xk&w^fI*{p`gx$ ~\ aZ왟FHZ895{.~ Ϯ>{$3wgܝ-"{g Ww(R>4oǸj"L2epNyZk[nuNJ4R g#{z픊:0bDG0p`A'R ̘73i Q ,LWn|Vӊ'<%+#wjdtJcNQ@Q\D {EpF`.#0ҤiNdXl{d` < kwv:\. H!cS2҈: !Uz&ii0 ZM%p[4U{\~}۵VsyrhK4oZ|6>nO%N9ܹvSж!qmSL>vͻq^ ѻ5Š4}6T-Xкޭ h#O?_n= :xGрXкޭ :%)m K\dל{5O'7# ʎn4dyfq_*bF1eXᏴ Bq7HWzVX'go?2Irp]D>x- Wqŏ@QKBTZ'ǫ ]#0" YyVӧf+5c-4~8^r)Џa]c0Yh6%Ҷۅ2 %j{zTwظIޖK٧j"*Tp p ^_͇9/I{O0( A*g\)sFOr!L%Xƭ xiA&Tz!/B%GjYɕ+/=س'݆_àeVo2|݉¸ŞAg9tڌv3^${ y8"x&Zzf_Z5q%8BitnsΘŹ +QȨ@ܤW7qw"y+[a/Fh#8ˢ?ik[K"NCˊN'VL)GMpi$CwǠ19=sۦjovt w& o9 ;88I&t _J퉯mT)*TIF9nW2@e2Oa&)nc|8|_t̂'p&8֔F!=?LI"On0߸k|4WS$1Z[Qk턴 jV9F[7QA-ɢ#)'UO)9v@iU6T[WrN%V6GUel9:U)8\67 A,~t7pο޳.Z),ɋu.BQ˙$S\$ɋݝ睒)ɋ-AZoI$/vA#5':/799DR>UrJ >aQ%.0xD1N7^EfnB̢V™Zr =2 -8$pRtQιzAp;T$&MhS>0AI`HR+sL2[U3:& Qi vy}GI-mvٙ= YIhO,nb)oF;@(m׌ٙ"]`T>19h>5琼r25Փz $F^z%w^9y3։FgD4#[6D@Wd d,-cd-Pu Kh0= aBV–ٗ҂,[֜ˠ`0`e10 OhquXWcيy^"lw*f'& Ӓ~B5HjEՁ!YDPh/a~=1[1`yl[.ooezzfEfcWLJbh,'(+x^]rȲk92Ԋ0[Ӽ/E U8Ǽ"Np8G9F0.gT?qie ;08m;<>b|9+rh6*G|t4ۣC6z QxﵳmL⑽]v[w'xssG3å2 zo>j[KQff]=Ȥw J}P%ealǐ̱=ˁk _rߝ!qH>茁bJ!^Yv)u_exղ8GACM1u%^w@!WZ@{OqP(yM]KK:/u׾ F8SU4&uᣘN$)p˅HRc`_$I[̠xz2DG^naffi#Ij]h-n$QVY<]Q#*F[7Ah:*Zf M2~i_@ >ןd"-fӝ:aFNϛO¦JGmXِNH6G?Ns߮qWV(3Ҵl2^Bv.BzxX۹4 ڇU 7Zavyug`mJS\˼ iS7@\,C=}%== =iw&/89I@|gU(U}+ySož-H?"|rr~hӔ|NjmhCj}~FH`ix 26p(i[nA>Ryďuº{H?1.ԁb4OBrVtQZNABP؟ߝp}}Own/NJfXOn>a'i)7'w觳ͻX2ONVtXќ~Tԕ X(bJv:YحHsǜ0oG? WR]gX (>o')&Î}kbԻ_G7" =DWoUEX]$r&_>Ru{(9L7OhZo{I>{wеզ n7o3;fCup'B4v{ϿmK2Mw?vxR>ZW`HTWR &jw\m;Xc cj?,ӭzah ^@~E!*.::Tul&>%JC:B w CMq7΄ϸrrAx!zL}>/K,"yVɌ\0WB"`~9Ҥ "h#OA®dLX6KLEjoC]2:Vr2",\ѯ>SVV#5tJԕBTD|]TYPEFqZla|2l; w.nF6,\ lbUSSP`.vD`͋iѴJ#feZ`X~r cJ G9,tpQ @ %9 >IIL _H ;%kc?;?Ѻ趦`6{}\HMbxl:I67Q]\бvk_yi|kߡU&1]aௗƫ%, M2x*_;UqЩdg{ǜԖYFx- |n׍'yXHeZڢr\s\@4́Ɂ˪ZZJ9iTz?.Z! u ~äP㝍RR@ޭI0 4Pb R9AMw %T#ZVdGLڳƝ9$$trˀf.:ˁKWlE9@{EXI6bJG܇ }fO/ڀ oR(!ڑ{K`QSs9% HP\B`4[lbۉ lӜd9<K jBcr}Qig-yXr RLOgKygvYSR(r$"Ĩ;>Ziex0ƣwh7 XjcirS.{-=kff@bA9iYuY;Q +j2R2(Vۤc {z0M꧌!R8EkŚޣV@^7,HGۧwcdմO/yJo=W/ <S 2 6"!0Fܮ>2J)јtrI[ь< dІ&~IפMGkVh?. rGSs<1dPӭ1w'2  ;m߶_X0ޟ) "N#n(5Ұ lkL !Щ>2/YG?&gxX?x_KM O~wr{i#ܞ=/eT|]PN<yȤ8`} p0` Yh7WW.>Ć@fާpЭiftѼ/eu7o.opo߾wYR33Y R=:Oo9Dp_D%cy]ptp+ЇKbwn3r8W'tk*L-}jbeQ3gqPEl  F.|E&|FO+OD&'!A9+A(D F8щiķ/p#ELFN.>K~1+h*X5ܭ7g!@*S]6lهN[o`駍;oI;y-1?l_ߤ7.lQXnv}z.J} ٧Y N; ?j9"vfPl"Z#Gxߢ_]FC*R:nF=5t1)Fa0!!Ebo?oFGGE5* ɺyDqb=(jzϖ OҢa_J!g *f́Y;+0`F*]hN89\q%p(VH Co_XhҚw@`D~2n/=FEkȟ4wbzڈi#Ŵ]O̻K^%2Z (=)֤Wyd.TT_PHZKKUlj SG&f޵5#Bt&f~r{'&la{JfMC='U@S .0 c./ʔR +Tw0|4& eo'EY9"h(BRKPH?[VV",JNb0OHd\¢8)(L abG{X0{٘3\SʯeeA&QX@3s*NpF59]08v/mNmΉlU/:.d8\S9Sh{WT Ig%i`jLk:+ؘ~J)u ؚThyÞX!עŒ̈́|{cKAgYҏ Ka~/5iϢNm?\^Ȯ4^AEr M}6LK *ydfofLo j2/ UIK9pd1.T|>}usр`9@&_<љۿzЌzs;m?d/`}!4Dsf&($_ok6w]jU +2}vhYԟRo;yjx]+@!r<a7+?QoX[J4V%)oAjc !hD5"\D*_=H4@YbHJ$3Zz&$V*ע|, k0541!KlLƂc ",FJ +)Iű[HQ`!!}#U# Ar #q~I : 6"EEaQy ak[@(~KLRR P@%UEFH"Vje[MFkhnMu28P4o .)1 #d8,Kf{Nj 4 MX5qŗ`q+mC)e"X9Eg͔,]yĄ%:[pKbӠ F ACˉWʑXPᖲsxQWw[x(ԟiknfa#dC{-ZTFFN}n-ql|:xIFLV/u` SU8` Nay+KJPma7&`i ªh9aU[􆷻m* Vݲ%d>3Fe-y/VGϲuSi׻2Gۍ5y{LO`/߾4=}55j_GiÜ~937CkL+regrZc|HȚoq.0vnT߽|tfv72.Qns-;0uv MoKKΖ{Ax}zFx-P+T){+~sq D%56ǣQ(I_ҒbJ w:9c'Ch[|fǢU)a?-Ιٚag-x>{Ia; ǎw.P>W= ͜,@) pVtK[0=s٥Jٙ!rd/[uVOl'-׈Cgŕg6)? #\:ֶAWh{@>3Y ՋTomyM!-MwtXbZp ^AemoF}%VPRhC ЉZe cQl6;Nzhs$?A`Hy4x_NQQwo~ys,IqsȞ|=omdl3vvS']ƿ9d6}^yZ= bH`o>h߬ (5nV}&^ ] .=`o6V4J7S8đY'|ز'~@M&__, ?[_: Ѣ,~0=e@KOY clVH6]0jR`q]ZMƷ1a,4o{]=iޱD5RvI$ l 9}e\8B+^}"D=r=8 EuK]@mCx 1j2!4kݦJ vq!=WE5aJ%ߨ*I \,/p,j1GŃ:?z˛~xsxfA[G̶rbMf7Opu+;iyDh%?_DE 1`xe|2.&Y:uu3F |NrjS%J-R\9kbz崱)SoB8Dd gh9C6<9o-s`!ȈEqSX0(,e?aĊkdKOL=id{G/CRF߄~[Twt7̋Qmn#}fy2 FHbH{;ENbJ` \r^0'+(#Fc.${PP6p>[,slއ:)1 mxJhz,CK*ᗷn@%BPn}IŚ4k5᫏' ""B40~D>(Y:_n1 Eb(6=()p( /HW:;ɁbN,Vs g9qBb<YEeS1 nR0}1stk |l%S@( 'v~ÿ!ǻI`yHH}љ^0ϟ%Ǡ4Gf^OD;j A5:h88C I0 fYnO)t=Jۜүq Kݼ"=Y J$e""455k{^R:Zq0ȨH{ҚHgASvslIN+ґM = vh4~મ_85SyT9C>++wI$iQ \ɒOs;\jweYqjəp P'r΍Pd[Ƹ8feHnѵB4*wGdW%NFP0*˜,0Ҥ,j `:<6JT;؜ ]&EwLNV\Pz vQ m#j!_<&F /ǭ,)/([ʕ*4-g0K $GXX[|}ŒܼEg:aJ4W xK&ʍ̍ɳѿt*Y)Ԧ؁dThM_qu55(Famxu/y{۔i`<PgAg s ^o3+\ d"E,f[x0RXZ9Y9So!zSiwd a[}%y?e"'*Mi^i4Eh;YMd N\3F[IjMܯ$:bD9 [,Xa9S,'>Cc BTyZ17&fRkoJt Ctm4a,r<]scWB`2pơ\k&H!rÅ$8C$'Nxx >5}ĔX4s~:!'ỉ=gv aMfDf>]=2bZ2 &y5>|eσ g4ͩ FK>w+'+ C: sH2R)F\i3qNAIBzZ@ezN f96Йf\YoDV҂?D2\RlAzXүׯӖ2Ƒ"pvTX|i-aO1.S<*0vDZ&A\K00L#8i'LJB-OLz")-Zʩt0OT0&M<!A :hCYD^>fyK5lsMBŢ(# w` 5v(0>1@5tگc vsӦTz:w|3_|X?C&RJ<j1ݧx$?:oFBלH~9P _mq&eHhf e&C[[Wj=Ne1F[| nwnfof(׳9GӹO/2|v؛`\MF#%K7Ml*5K!ɇL^?=6Y4#-7of,&Գ O[pRA57/&1L}@Ebh-}R4ΔN. ٙ>a4{nC7HLSFﷂoWv뱙=qlpO}eCf935? ´eYKuC:ʼەq[X$ݥyIwIoK{Ϲ6R/uwm:z(nFK41z>AV6y+ "k5^cM<[~2^A2GˁC& ^՘QLcsh[@Y1J7h﹕Px/ w@D! C `-<9}!"o!cdMf-?\@:%4/gYc-o6%T"H󶐄™܁kheTsc6` LlZѵ[R/?_k1!AjT>gpOuR@!sr¹UE=A\($1DFgw۬=wJۋ6 J^rY98u4 ECpQ:1J Gx.uvKimAHP+#-r $ u9zUkz/f#7a6vv:)Zx: ̇ gY=QYFvC0Ȑ,)W)[kσ}sdFXVM?OopQLpW&tlN_kRiAݻyoebgyζ.ڠ*ҬN>pq/n4WE6.4Xy$}*%Yԅe߹w*BYAS6IB To>Q@$>SC8\WGΨZA Eucom-X C6HA ɃFڍ m |hWVjA&N3Sf<.syOiV+Nlg.>`j!4J0灍hQh X'2}Z*B_d&{9cAW0xr{]=ŃxX3ήߜ ?_ }Oa9_*V*xTUuީxRHcHeU b=2}O;*>iG`Ɠ_LL?3dZWrzUAj'|+MvPa4F#c*gU\x);md#3`3˥F#,4b"tZ1uMClHN JЪ' ~*~[LXJR'PornhPL >SpTPgf] i% ی9;REunpS$9h9uPO|QwDo:\kQ+~\׾7fVeptUE=(WܱTLs04üq rM $\ @BjmoC$誸kw&az?1MRBf,rLK IZ}!̅.}"4apX>ʁK^V$ (@4nEUنo ױHi-B2Jnz'5PxuGX  Qu/rh ~Tkƚ]ǕJߣ=d4Ҋr+/O%N89uc׌ ݴ ΋̺{X1s}%WJ;DP j=f闀MQ/Rd1/ SwSRgcLBQk.c]y;]REJ۲Z *:Ts0q§_ pۭ? x:MTp?F8YYYoC't +{TRs({{[.Y.NK?NȞnl!~lM*dxZ6Qlf<=LXcm(qd9N5Quz7^ލ<5%ujP8ݺUC5w?h}h8=?xt tt+#.=5;]6DJ.W龯M+3V˽60M^K8 %Nkݑ鎮|N9]5ÇWb{㵍fʧ]"Rɖ9Eս6;vGHd]貓m7< [tnnwE*=~i#u7l&aD't>XRJJo`GceN"b}"p b^䉓}d&\&0EFS[R0цɵu:g2 !nN؟Q.r,MT(HsD8E.S]09(|&dPRR3Z+-Tij^a 7/MQ1| ;5NoR V.*U_0W93^nfo?1?>ݗ;rTbKXJJ*Pm귎G=Ct ?eZ{ȱ_e1#G$ƠO0XU,[d҃{Y%%KXoʨn#H,AN%{k޿!D%Y%*U@$GBR !TIIibd-BBFN'&6ilIDbFS@Hxb)&MAbq6\9UpNgB}su,׮w{^G([r&ƫ:݌ꫫpda¥I )^ayT>9 } hvb-x_6b7ObW!UU(kT_jul+^`S=4ªe{hLPk@ $䜶ܷ{$ x0 BD .ݦ bZ\!Mq˖YA0} }qwM1HE{x鑃].A:ksvLTF{O~fnm~|%.Wzqge1?‘LFgwy 6ʿd- nšI  vNA ^pCJ~NqK}KlJb>r9뺌}[: ctԺKw޿W;N-&D)=JЅG3p>a}7J"SwWMoSp=DqQD> rq64Bxَ2g~}ϡe 9ї)<ݼys,BՇOV\ 7Ԁ|-V[GDD*2R1 3c#u(1Y(ß3cjb%RGq,S!f9ybBFm #%n9)BɽZc}j屁XI.lXXY.ȉ3k[}yފTIy/6i8_]a /Uhq S{xIP#* E$))c6&Ip*00c8%1n~A|,f{oo̭2;k{^u8lW]ǚO)(Dpߦ݁KeG JuY{PsF T ..(xԉKy,q8G@ѺPͮy!y.oj ayX̔FId%M pI0K)11xF15A-lEM!8IQ2D"v[(A:LƁqL"-ݳEcv&).O .-wI=<]k^r ]B!\N~Tr [2cы`1sm&b MXe&gK?Wuڧ O_e G;GLo?~Yw/0jDa! g@5M>iQCEv(Gy#kqMډ]Ղ&S`nEF(8 ͯL)oݡAJnmu^ׯx6`_== ϛuwۥL{xm7".4Syxʯ[y|L*߇?4p%Cߝ7\eԬ忘 Q`5c^Xj0/|>wle΍˂ |T`NDGyqD)en*qô _ 1&]/NfͭxxVxçOz-r g&!ڦH%ErOVDP*9bQ_|'l^}QrϖIp\$F8D\DJa[-NEQd"ĔN1F"Sˑe!`]fW_ ~=An:fBp$%aaKO!Ǭ&@VKot;LFëN^FTwJ1*\Fܸ@}EhGLZ04QP|[ -ܖbO8=_sDgIijZZԿ)L!Bj_b|/cE 3Jq݋]VXR4|K!Mh,Lxim冦/_ׯFbt%>}2)&^B pFD!Kϗ&|<:o=*fX*~ӔC_2\=oďd bCˆa6.*Va2:!hZָ>5C DO#d5&!fۗbD4%A }MhA_Y}~!r@3nԊ{?Cj+YY<+pAdIX5V8| w\, FaB>Y_>`5G Q"JRw^֝nym/NYA&u4hdF596QyH#YgF3pM ,x_U!(w\npԽVi*KQB"e ؂m'{kZkMp(}ի*$ԝuZ Rw6qU-ug{.gB"B bZLbGFKj Pf(^@B)t\J#kYMn ybbƆDD*f4Z{xuڭ{88=ZqAY,ew\w\=2 'mG]+g 񃇐61_\;̹Č%kD[X-wOCd *YVo(lA %VPS萟Z?%ŭ-z *,V#ZKyՏjHtJУҘQN -v@[nEF&xKȄO6%6uڬ%"6Zu[t7[DfV"E^6&}5'Vk5iywRշbR":V(l,"LЄp_])8yOA~CT\Eعy6ebWOxv4#)/飼 hc打x=uvNVe(;&*bݛŪ'tJuL)+} ڟ)WAG]SԶj e\|?Oڏ$:&lJd $!}ص&EyP XET5r5Y U^Tk#g6"|v&UESgD6"daNZ"k8.QdwQp6Gu~8Q!ZS pY)9Tdɮ??7xUL42J:Z]^j'kMu 6aVH 7 tE> /M1lVCVbaǬZVŰIS:#iؤDlO,`PRi"#dgل2 YHQ0s'c:.\tuos:ӳ)DY!R1$ #P&&Gi*8U`c*x J⒘tvT(iAe;'WR O. iHDX1"5)7L"G*4J#-G#PT!"l%%DDDŽ\H)lYl9Lő$#Ra,R:8* 醝\d67<>:Iyv4%m9`"q˦(yc0~4>O޿~~U(ʾf6bj_n"bxX> rPI\"<~F~jbگ9we޵6r#Hb ۻ88əL}ױvmG.߷غ;t2%J]bHsC4XjHSh8Ũi1F6v)CB㎙=5l# ڰfxaWGQ;o6'װ۞c@dDhÐb)_νUrHeB.A$ F=9_ѱ`F.,#ŔsWC˳¹3;AYT5Q<` (\6ܺVT(V_=OM+MZ?>K&BfRUAQ^d9|~{Z-&CGI(&Cz"P.t t 4_W_/E&=G-J Y;5Oμ>Y':AA(0@ 8!HD`=ex|c\g Lf4Lןܸ[5[7X(cLYyߦ˻Yߢ*hs_<(KՈI2w ֣0D BG$>H3Mt966U@bSv ,\ = gy=0J>ȉ t]e5G%F:$Lb‰jֈDA8!hCN$B9#űǑc 3G@e NLUd9 kJC`p[^/,|,6uˍ|V/&A"yz; !l|\S M''܌+֨}z{yFSJ|FtS.B x+@#>L ״ܞX% D(DrpDߞb/өq'HIx_jz=ڴ3>T;:yς/xzd&(_ߦF>te(|]zuFa] )ؙxRy EmO7{[-o;Xz_x 3FIƏ|2-WoELr6xΒa0V5iLROcZjڜBvv| Ҁ!,aH &fm!3in'y-xAs,`g2{VB SH^Xr!FlDvë^ݙUQauU HKH28¡멀B]W$XSu8]JJ® zg>رʷE9nn>C'!uŬ4]\dgn7n[6묍V/fm>RmR6eNblvCD9t'tt\O-W$̱dNNܭH;pעkz;HD_#J/lEI}Ȫ^|-X9y؏qSI~^VH}"ۦRI w(rh֍yZKnQQYKx,aiC MÖVcL(n\z6g^5JGaN k9rVmגvD['e![qu2.F!4ܤ#i,QwOH37̥4 +;̍˕:x?eH..UAeH 1 l,rcwQQdE*G#DN&cmxny)II:3W<>;x:8|}T, kgn𾹶-ln|Zks;<ϫ"{hǟ>0.ӯ/0tRv t9 Q_͑SeAJ6iva/ɏ3Ԁ,;H{܉Hij[ &dbWZ)(Ke3Ny}3`Le'/ϧɷmrӿ,W>R: ap>*FIXBh|9R1C{UDKu2tg'|6yƕђ~TT7v*o5f4'!$?Ν.yntlY{˞j~N}t̼PO~>8d̑N-z yeyCIBN?G=״:t( יz`~97KUulj-@ i0~dUn |:1"Ӑ`2jX@E>LJ J"qXJa{7~~YZt=ڻU}Ѝg-< s@]L-C.iT5Xj)[J4Z)><^-y{/r۸Jb f ڬ[{ڥH|÷4H pyiHno8}~Da֩.ӏwrGX|lq;w0!t~_,Ͷa퍀dѬw]#.V\#6X9:s TDbA#ɌZcڍ0'TJI`3uN_?ߖaͰB# ]PɾSYPZ)3ơNgp/wBzt#M{m3CΗ)= v̒I ؜ׂs|`_e;BCHwIJM 5ɐ!A(a@\gl6At:+b̮Hs&"Nl(GĤqsSXb oxoG)[ sEaW6:iaB(ďCbt!z3H9 es.'G"f#_Y>G2 Lf=9xCME%z_9X8XSrM˾P ؃Ll$lsw;s/Oª_nRt2DXf_W:nVޫ?_lM/ME>FGNP؀9΂7F#A15Hy Ad #'1:,Ng{Ყa*蓼^;oE =Q̈}loDnD  :jvhV-M>a9p<|yJpS؇әi-C!Ҳ=f]=XWtW.;zvyyG_ZR֢wx]YYKKCuȶ>ŤdeE WYRU{~.\{upW+j+VN_\ڋ}Ba%\-|yI8] 6S75\5h+sFa/9\{[8_6 |z`2>R.m%2+գ Qg'/7rVa%+GgkLz({ks>> 6yƋG1] t߭ y-^G1 .mF[ͼ5%eTޖҐ׃*8|WhP*?(aL_>B}O`/Ar Kp>3=lG @tQI~YUl44CpC8IsoKˤ'P#H䅠8"-䪸A henA+bŃ&3_CY#~\*/DQ.o]؂ZWpb5 q0"`2FIB9GVZakGn@ŵDMa6d|k~)|4oq%Kwp`λct&{cc{L2\[ |(c<( ,񹣆\9$aKKB0cVKu )k+铔V[(R4CId/5Am9Yޤ͕u1rnz)=8Вd iPpZp_Mܡ.7%#C p؂XC&C! Y& {)LI`2ĔtA/̈́j4zZh$PLyƿz99'O9:_5x2۩zn]LZ2`]9my1}^Z|zX,7<o!Plv:ؼ\Z] 0J>[NlRd}MYUN$?⽰旟X\KYIe4gBG%*3Hpt WxϝWϟNj%Y;h$BE{חRY y{5~R|nEA "Q6CUKxI<H?*uiR,8._qr/ BK.Zc. fHf #E"bk췕~7j;0p@6u4k˅?Eg\:{n+m?GB[7{X3ړA=/(-Aq,P~+:ZB0%ieǣTvc ?6"($^Lu@)2.Dw8HSUAZͱ*lДk!ڹFB:&L#v\?%uN%#1\^HksPB@b` Bn׳ȝSC6%*ע0(briȕ13$b\!w((ԡuC3pݐ/-TwwCol1*`)g;w)ER!Jp*ₜ ~:K_g|%ҤGvFX R_Z$2QWvꀒ.+1Ht44ՍBmݺH%RT ~JQ/~N4A]KAEʠ pN4461L`cL͵'0ID[Z{x0SC%31b3FAU5l/:,*Μ U1.+(#HpR+BYw))3 J01Ϝ0F™}Q~b#n|p@!IGfQj ެrKG\MIӳ1*yVZS~nٖoqh=Y.#^+[k<+'α|^IV Oa8X9o| БCdze/$٨ f\!r0@N(TNs?y!f8By)h8^"S{)R7,MF1dtEq.;qJ[6<19[A}cO9'YiiX-eQaKdML[ D{6wӎ6w-F+iTS4\/.IW 9[t\3[Cl߾y=|6""(A{,Р#O|x9-=QD3!7ZΚۯ|lz!ݍ~72ӣG=iZ` TB58>aԢB HP|sΨP(v]u.fe;՗w+.[H(bM.:?o;yewېi_21YH2o7>cݗ&M@SЩxڞåZlߔiSai=~QQkKhMUTQ [*1&qk~)һa!߸&]w{Ǟ»bc:hNJ`ݛwK?-ڰoD)ռLž v0^P52"J1DI%b!. (9?/s Nrf kϙ-JHf*yQ@9Є/H0IhV> u%]u[F : C)d4?CAcw9-jcoYXm+ ^,,ݲF (DWjD\srr~Q`֒!LiǸ%|UBuq %=Kt޵m<6Ѩ bgsT|–؛S~&CAfp{u]GaU%uz+fg|k3j3j]5`'Nʗڝ /_5(nN$YC8⧳-U"\$T %F1wֺ<#5B͘\~`ew0dȷəQʱK4K°&aY3˼ tA,cy,h[Xb!BXYDʨ :`,xj1Y7Uw9+9]+XJL<9!FÄUi![1JE>o[5 iew\ ;{oq҄놘PKFO\$$eDÜHd<_7 hZ<}/wfz^>xH 4h /6IU=\*jI IJ:c&Xhǩ"T*CAۯQAəvprmP=s^.8kGR/z\Z1gHTs4@Hw$ XGJ mo%oZ]a(Bt>fw5X, ~va~~ \hi[OWן6DM-3~?r lg#AM=8xY4l2 l4xe_lU܆INןxi|^ZD*sHrBkqW},"9;Lfi J۱ݛov;e(3m"*uIԲ.9WϊSUUbKxZjJȵ+nFpf D @;7;\8ӈvxN3dN ‰Tk@r 49uԧ3-:7*%FKd đuڃؼa){)չ#p ,(J1@< Q!4Zūͩy=^yAcix9+M̋daP4c<pn.zχd{.bl +ơ.n4bZQPBX.*X8h|q˥ `MʡoѣEz|fo&l;C/_OEqР ԉn.ꥀC#]_/+a(+2ϞG-޽pZͦeVş#7me$Z|QqFpU@/Ӂ@!Hzp'͚{ʨPhnciHtrxҞVo颂C>-*8P_)uptQqMxyFHf쳜Ovt[A R4eZmVs 'cʩra0i!pUg0WME-X+D\4ZnwwGWgpYc;KyzW)GCb\*!fEEacc@ZN'OWkjts]|DOYĹPW'f+V^ddn5.ke5s.64#ॱ|AͲÁEZZc ~{]HeSXkDŽ^M 1PjFJt{Q ߮ XeD[e?>l1p1̾Vw;5! lu܈ŧLT裡L{ ϻv(=0KQ% XOB;`}BT{-4Ye=o8!c6p:z}μN^v o,z/ښ4QS!>mL)?2@( g;P/B6px} Ԙ(VpR Fpy nZ k3./ W*]Lq"bL&3z-LI_b6km_dv痹)y3՞^J9TRŤs̺\1DjMnKU)GY*F7(öhשaW t,OQ#J2.\Y.SC t,gs\PwIr)-I˃NʸgI"A"*)ga+ɦT.R"éJ8u^:*fκhy-o-̎eB~J Paz^Z r?>~]"JWo_~6yؗ+dBXRN?{v:8㗛+&#B-V:L<>b& ӣ(1J7\yv)K_QRtm8 Xy}uq>*A6ID--"B3djI#ߏ5j5JE='ʦϦum/ݰ-/cO_A}]2g;⇣EjQqwsuvz.f;ކy]WuMW&]}&6BMkܔ`t %8vJأ ׂh0dUi,t*I[0{Fґm&y4Eis9:~=o4 MWHÑt!luPp*Un#{oBd®ԁmwɺ*~^j|G: ~5dGzX <<ڭ3qAWE$ @+TM#DUݨ~k`ъ$5BII\fB̈jtUƮRF+ XLpx ثkS07K xrOtG!^[,/tեO"R$aUSEOT1AMQo,4 vB&B>0d={fGe N7);=)RWXWCS{Bh ?9J8' r*qWYw\n']=5%Y|?! .M!!)OU0|TH 5Yڟ_R?7_N\o5]\ ^b؏轆A>nZnb3ڔ+UznVxS+v2}%Ul}W]>5E=Dj̩-)"W;Ŀ3oQ*>/wi'.Nz4BэOJThԆS0ZMHب4|(jCE'P|G.lG}u-],UѢIZe3IW^j z6A*pj E"J`* K7mI{ˁxҠ~K-U}4˹Qq @cZ;I,\/n-R[o Ʀ8=gF=Z7L=6\f a+%>nyHz889#ZiFyQfIƥW).u)*#S刿˚:1u AzY~uBYoP#$dE1.9saLJ-4,gw|L󩽶/$M|iFH0D'uezU뭠Ȧʣ177Y}ko{g։IAHH5{S R3w{)7K+y!1l>D|:ȃ@+@tJNR~_ee eh'Zh 4Ox85-J!+]_'.wjfx@#mxt}q?e2"% iDTNPIKQ>#-߮ _o a}Om!;`SZLՑΪȽ,2b~7b%fb?__~6y؛+ ’O~Z|R~yG+ ?wglj>חGl_tpoKboOn~:ƬjγO_qϗj*jc/هU}7\) .| !}Z4Ui`1Jog ;i7!J4UJJsVDn W2QnP("t* - ܪqLGJ˜g 2eZ;eH m vz> R`'AO=(YdKJ.%\tT)<$q C4JF S#!XMRp؏Y7L&Jv]i?CɄJ ]u9ż)%bbwaC =% +aGL(QaS%̢N0k19`s8Rm[D4~,=ZTCAlOXWVX-9ݹ] DRNqk@ ,1/bJM?_hݺy 3l>qt} c`L$P6ҔIO\|Wf'|!Htu:r~vS " q*΀2EZR!fj( \&H!- #a"d Ppb|a([xGOݴW{EUVU/.?]+tIhu .TzP]ɢ>B }C^,gcѱYAGURwK ܻ;%nG+#^.Wz>=v G ho%AN.qvvo,5*|noEw0To֢R=A"MQs[nptZt:UhȪڴB-k]'TkjT|/qԟʞQs$m>m#=#qUJ"9*_J✔"gE]JfltPkKIT}e&0e'x>ogOl1DMi +v4ʟgj1i)dU3,Zr#?ߖ" f(qpU =izOh$[u:t{D=aŮqX'!d&4exB Di)D ]aJWL^w17['S\Wcǜc-Bk1H4FLFm a ݞ+I<~dZ B-5R20(y4>f`Y0(TxosU^WTrNקGw?Aĵ'dHDnF߃,zXK8RDꯑetU|n@R\~YcDKArY9 W\6k\ڿO'=ёk6fE{pBkwV))oQ'Dߡ>[<}DA2F`XGHἻn/ouo=Bpo`ނzݍEc1U:kzmP-%yNN/fm WoJY[!F-轉[}(hfI}1toGހQUw,F/!v9b=U8Q6YLaKŨ\*G5EeCU27 fR. 9Y]U;* Dפ"&;p$2 ƴk`b&q:Tևj8IcVQxeRir|*vsu\1b,Kj\s[FU裥Iܶ:]s= LtQޅ{I/3j{p\f PP:xa7HW*ULQ^!]ӍPH y/CA^RUxP#*j> @uO TmZq]v# ަ&S&cWյWH qI'to(,rohI[7v rͷ9"GoEWk(_:RN[H\A9y^|Kvv=lj- WD^Y,:Wt>_hKNBx;_HaIձh^kNr}jMJg%de-lƍG[8JoZݬaҎw-?kz:R-6XuLkq}oޜ}akEu/:rYo?iILmjk?5r̈́{SlE\ߦ}T+Cɝ{/=EWҳc\M7V}J&[7y-&]u[,yusCͻA}N"XbzF r?o=bl^to[]䠺v9On\iOM 3O/R GK&RZmP_;Vz=}?'/W>PϨ)jLc6SK.qKyTe0e'xz%&ov/g ޞ䭽R5.@w#5q!tD[mybkUF+I*O=S.I˫w7P[~܎onL#;mjE'Mz-`6eREM{j5KqF;@iU&)ߒ4JL=:dƁ3 8I<'!nÛ8Sq#mw#g96i,ۍ#ʚ&d2-7̍q޹k#MR+5f:)鮌,Y,Ϯu28TqNV5Hm&9sr]>0h|ޒAw`]g ^l97gޙw1p^`;7}#46WeMWӛ76$SO٨Ԑ /Nq#u y',A<4DATgP{C0>W; Ap蕙!bޔ( qk wLgP[|I_25#~'a+1]ykwpőر"͸@!胋aY \ m % 3CD)#Z&#Cl;B@ kzeîv D ;C!#H/^7sJ/_MiHfP}n!7lļ[<(טS&|CѽE^.~Zi-V˧gMĶ(ϫZO= o.԰nN uIbAyBK#~"FB&Q 9IHeRi,bBA"c"SI|Hʁic0Xp#$E6j.HE (9F! q(0`G@rC Dt"nƾnuzXVT毫/틃;EKN2(A& `Ji2BC'$d !Tr PP^Hu Po#Yiё8A%9V2\l|0Ht ib  V&30 HJI ELhD0 PX"0( q@2J>-{G?Ui+Ӵ1mmQ)j͡U|˗ݑCw@_>)X_~}*3 jף?g7j4ٖOLU.[~feb | D-I'$hoR{EOͿ/%UoA#&cB^MRMێ8Hgy' )Qp y^u:]j4WK2RifヒE@+CUsS\(Y06dq^{Woj2pշ^̬fm %Ajڭhچ*[E"!}@4"k:(.g_vC"D9Ow/sLUfd|7"ѧUT+Ӻ{T 0#Tec$lLB5ʊ ! 9yQ,mIԞ@B$[:]=pԌ#%Hxci a<_"$t@ $Hmt$6$qՑxz(b)K 74@p@&e%(2~ W_*#D&<$ To qq#!hB)"'ZJ2t%KI{ WlN/?d~؉-]l#z%=82G\C%^fyK=dг'u^-Hw>z>U->BlY:#6#U^}k秷+.7Y*eDG-†_1u>?|'Ot^|v79*/-mTI;͜@e wq=.9l8.lyTu"jŸZ@q?WʃɃZt1"9tMkF BK<.:q*ZeFˡP7Ц} $9ҭyݡQoF].su3JN :6-̉*q'8)zwQӗrOVz}}R9 !UЦܐW*vNx;Ym a`tzj+8[0fs+yU; tuS)J)zL* v-ՠ;-=UWcCHM\7r*EXE$SoX#ZX_]Z/h:?>f:|F[$;`ht@32 cZIqQa cP)140(p w i gOXpl}-pʨXy꾦@ !Jd<`{1޼)#Hk:`0VgZq򗽭VHcn܇2l2#<3T[zlC+UؖI?AÓ]KW2X˨^.֥n2)d4Ov\(ndei3mF VU[4{\b體¬j)m>Cѐc ;Yr^OnBvBg#r;sAؕmTf|YR)U|.03Sx.u&%s\yns:[dHL96.xdIvdmG7RhQUyMYtIGKמA!"b_j%x'T=?/ۧT msygF@TyϷݩx#ELc,?ީӇKFL ~{wS.%lXG>Ql xOAߵ298*0`.fGt1(Tt'7VUJUn9Qo˞؁$AXGCU%߰Q5@x>?|~ A/_Tx5B6$bV`\{Cһ6O3BysS8Lq܁*@77|-e}Nfl>LYBn/g#1Ӣs+~,>&y7#1/[Cw2YJ;}a|a6-ȯϦh$zIoԄ$֟r\5}_}GЏJ "XƄP*[b ]_WS* `jVeѠjc *`>2pcߒ+O1v^M_I[O] pnFY{[)ehukU7.4OrNFK=h)%AcX}I׊ $Rƻ_?6[uRJ!X YcWȠ& mAyM8{Z Gm<#,m~}ǽ& SpN*C"wE!\Y-`Sgh@R ! e zr_ʂeS(DAF( 2R<,emQ /hrt/n`wדţ'O y X1vK4p@8HנbAEcEG 6weKj$X⼾UJ[M/ džT[h/p@ӯLX/%5"ȥ15` ^!}JƚmRZZy*cZ_HmY4:, SM1|>ShoR24/rsڡtJMH'jrz[6cwARK{3rAx[X,d Oy@F:T a Fj7H:+=ai \*$4nBUʟH_I]\xp`-$COֳC{TV6(Hf@C{Ԡ@X۫PIF $) ί5?.3gr2'R)7kRREXss.s@s >YK猎t9HmS 3BuB" _LF'g3®)m+tJY&u)4&/MVmW mG\kҫ^UW^b#ZQOhryg,1Y Ag=R YH`uͭ'{>>~b1y^z^SZ&:v;v.?;R:Z06v# gP7rz_j[۫KXOlMy y'ݬ"46VY2]1;]c-v=hɃlQX-Zng?j3ž&yY0&-紴iKǪ V&'_bX5K{Е.nL oC$TO dz_#<&D-Q8 6/QW#D-YDGF~AڮmRLU8Iт诽NjJNrLc.4G2}:|NpfT!H~doMHV,uF1ms7*5#Gi~lpG7B蓒F:Ni+믊Qt h^T46MPKdMb7hMzc(+Wp1x ApF4^Q5=th{UG}8)8bT} (zvs= SWfՂ+DZuO۽Z-B ot?<#-,bCd+dwlhMֵ'CحP:C%%B^P;Pn*#[Fdp- H5-6t螡N]iau aL1KnҺrbE$)b%N$+9~ٚ@+:G-{MF5e,L4q>:ԉ̪;sؒr9 Y#_&2ɭ)Jb5z3W +hSNs|ͥa2KpJq][B ]K*$ /JvRI\+YR cQrcXU(/S1,RhXxdXxCbȝ?vizu`X9eɀ6=WdFq$m D'&IA<{0c&3}h#x2TԮkH@)؛(B4;ȠOCGeWBa@j'7wu@QBQڽ"%N²N,IrOr+Y6>Mf ~f4 ?^*xAhCJD]F^,rN`}CmO/)8zgjnsziѪFtڃ:<>dG&b );#z=5D } |pO,>uyZ]'n;GCԦU]B4-d1 gLnfT-dO:kԅF!dg4=U[Ҝ68BCWZF-FM?-F~h]?{2slt}X> hBU* [y ߗ"T Hw%f{\?F]L>$$cdvtB"ƒBW^T9Ey<:nfoFTKZkI m3b2CaRPMmlxm"o\5t$T6=!ZFk-]ٴh+VfeO/~?Gui[iZކnMPDU]>K>ۍ\N"!g)СU12WWhyr3m&Ss"Ma2ml%mwa6 -C2.2[ZD &&W!d?,ABϡYR?$:nCDYg5fٗO΢a=m7=iZ>l]f`~%>#N3@X"Y42{+qY`gOrbc*p@nm9P-ȎCNq w uy _ysݨ+wW]j~ IP.k|×෽ar @x>%JM+$^B{*&9hMjrK P2IA_8R,SY,0 6 Goy؞b;i[v[_-]Ud5kpe[as& 1A(o@ʳ_4kW yxvl\1I54˫شj#V_ίSmnupamXeG :.F q"FjN|2^%\"':y܋Ύ@'ƤSްrqFw ݆x~BMvB+^< Js9 iXj0).*zbʫ5xugrChwm=Ml Fh5M>ƚA֌Yk(KD4hۆxyDږ!t{&Ígvh}y]xZ ,8 ! n6Qa:vRy a~LehSPfNZ/7!jlbp:C>YC . 9q$sMqTvR=0@ 0 SڍMRyJXƠP$P@Q|J{YVRvj#bv[Z,mU?ή>"D1סͧg|_Wy&%FzܭP\5s^={O9pt_/.8%)ɸFrM(Etمyj{{R[@'@QR!H4n=FThbZ{GJ W2ՍުwSP,Ƽ kTp]3>\ rV2=48Ts 1cOF@YMA D)e)CSIoGvho ޚdL:N!"B$ =Q4)G5BVfL-H$;2srA_SDF25jQ`#MG|r6EvV#BFמ X #I֞ P0cdZG᭥-]k+֮iw6X/(VaIkvHVvE}[qv[]>wm}_Jyty4*s):s4(X vZ78#}߯\f[w}Sh^^"EkD.XY xp%Ѯ!TޑmwWotn~?P;03تo}vN# wչS 8PhS;RnZ;\*,U`eqqBZ"0uSd.+䵏kwE{{룪JpiQ:ab^%QH;lm%&1 EK!4.} o{aڪ{tJ=ыr2(dvL>b=PEwPVk _*R^`򂳀iɂb1ZL#A*|q0iVٱVUWuw?:]Aq2:C?ux$U8U89G@UO޶#z^|`F*8a0}uSz%Ǜ⏭{>}O\-m]e߶k쿍oB’rI)-`hWJ[vV~<3_sHtSx 5+<),yuSV:WO0XzK4w,o^\u]6⻛(z?D]:- ۯPbɆ!d߲6wVl~7-0:>*aylŔQ`3҇d5lM4fLkVn/Ao=Wח>8KP]˭N_ޝ?z['nZ=:)^Zr)2SZ{7^ jIJ%z)}S㦥m,ڮW˳)pSn„oz~^ΟyBva4 6y /n7_' ~_y~",$bzG%oC@jw2DF@ TNЄucagnڙdϻ|}U{Y"PC -|Fi hUg(LZK&҆YG+Pc(ǘt ˠv{4W1טgϨ9`鉩2) ?=pV. gYB;Q%2Fhs*(IDlLSHNOXWҷ:hݺ:[ןJgyc3P +3ͧgQLG^7?m9,%Hҷԫߛ/6^|Ϋg8z>_D+Ս₍aw`i8V.O/=^2V 8Жgo"ئgjC)Qa[njR: 8O(^L nn'TwJ{^@b85W2i}qr(>=~7-9^w .zb2gMclƒHOY)'E&Hhxuj3!n0X <0 DͲ/ e7n%e~IZ1kllYиgY $ "4VKJPSnJIF6ʁ˜j,48nfSJ;>`‘J71t&ض XIXhR݀::x,FaK7/68'԰3ȃ\y'ȥA t9reiK6 C9hEAqP.efO])VJN̚?\cmݴ-M)pW7pi/ke KAVYGj׿mXZ%C[BzY*ZaQe lH& `a!C;.jEl28#u pR6] 86p5L쵠@wߒ0wO,6d35;HƌÑ6$b{vzHGFH˼q/-"#?:,@*ZdB)LQXX?2pCW+cyη:f¶K7.uIkcjxO" ZNXk$: JvCء&: (8y&'gJK!\8MP#+ՁxL~+zߗvƌnh iK2 z<KIBT mJ9*SM%4! 𾑾! &(眖"QiI`-cR`}oJQCz-L)@>IO9)TO]ѥ xdw ʄByUjKi< 2 *U/k-NP/iuND[Cp_ +`֢E3o}$|@( 3ØgR7mb1E69!+('8W_y6IL(,ߪpw ՘,f4݁7ѢE kɬ|q>8W (FC7񇚤Pc=o}zL(ZJv%&(Je(KAevb ArY:&!@&*0GUFﺇrL #qV.G9Kgw=S9x}dtDtHUk :ܝ;ӂj ,8#D![g-☝jkxYI^ vD'iX2ebMVoaNz"rbIR,25 Rb* R9C 2"$MBE'iJ_bHP:N_,ŬQI8i"%HAhe:҆aUéȈO 5((I#2y&"I,c<ԆVhKmA[ CP`y]`0:u*gai$B2dVxu@|9gYJN>oL(y|* 64am33I,[@ N` > qE@-?!kʮ!)J}!RPuzޕ!W# :d=yh'n~JP i, -&c{wNrNVݱe/Z.swٗLP5fe7 @e4&e]\X&V 7*mXbI!\[\W6r47^?ʠPDj#m1B"k,F@>m֒lPDItKrɅ@`=/Rb ).6 *>C'9he#e(xG} `nӏzXߙ@)x샛Qd wS;. 8?&;ǶBԢ\֜<ΏkKRwEI*؊B /'lj kK?m=`&n9Y-<)oF8_ . ЫPu]jH1vHdFJ< 5}55ޣH 3e{rY֊?>~acF9%tG:eS{p˜>]F.V2W))\%(2ր k,lxɛo gɬS y88Cw7ӣmCv6R<9b ,%Ń;=:sٳPqNr99=:Iq߯?_Y}tIHY&_ —V_\|87}`o},}Ӭ4F2nfK$cz3D{)dlG/՗:%s7qt߀Vǥ6zØڽc%0un|1*QkOtQHxX锢Dxve8p \Fjr.KȭwItΔQF>E\?Q%B SuЁ$\d鈄#^Iz }Ɏ~p? 5,.$`eȑ9}'+}\:+"zqC7~vЋ{!}pڑ\+&;=#ϹP>(row cv簶' 7ι}Suv\+铀q?ɥTQhYMۢBLxnY8og=jwhRT)~7 I~B `C5|!B,X:rK,B%cJAx!hrEJytZXl83کR̅Hk@f!*Ŝ+'Ǵ|nv [=Q %i *T UH9Kyb%xE'>J/ wNY2FhB )ב~Jʽ 8夲@K]i^-EVE1KBFX;ԚiX ߂3&,ic-!hL#BKvTȨ"w)QXdauX@,0"hЯyމ4dp0!SqT{@ucʪHxP(-a 4u"f0u O&`U"%Q{*5 *La^5Vؤ6 Z¢T{8biY`(IP)GS z .&@:):<_Ah+1P!ɖk9EыH5fH03̙a̳>* @ f76RBD?6qx&*TW/'K87Ǎ=7>jɺ.j3r=#JO|*g-DY(r&P|)9WTU  {]uf '}ܥs[ q]@MubdC~携w|͘ ZO|Jl t*lgL~yxWվz釋?\ .u((F8ߦ^cez??Ӷ󯽹tX WOmL_f+aV*3mvD^C/ MЙT1.*OJ[R`zO4۸ʋ71hD9?\N"~-deGOFz8"rHFۿmHygm,?:rOX?"5&}6Z=-zf&YGۗ3o!ҝ7P: 6QL#t%!&SR]$%ql^Zt@Oa\@>W}3 MIHR`+ w&T#:(qsW>i̕S6g^=ɮ3CLF&(Mjx<=vB񾲖@#jk?=&b|-: %b@N)*EH`<ڰƓݦ25geO9QJFшS=3mv cs~-ŇV7hBI_opp[RzKK)";8-Fv>Wf9i_!LYs"GjȖݧCYb;{r$׾Pe'Z#y+=X7R`8!vl;~9Y <),/`NŘ j- ,2;sof!p reN4ÿߧF:sZONnAK<,(AR7O ˔L ˔l& ,C#('Sѝz9 54 nӗ6;q%xwE$V}tJ+uSEnհ*jQomQKJXQ BUZ%>8,N݈9:8\-/Kq5k) >37~;jn>]c~LjV٫[ Tu`[nSYKt%fʭ[.`Bcg !wJyH:>S<]0mLbHrlLkt0KX=A% -z)Ž0O3/zi,Wq9؎w(X8DF#$=>#2i_L4=arN@sxΉ1}st8QN[ΩMK^<4GA31 1 }(xPWO4@۾[Za]uv1qZض5F –_pJX}\{ils 1+AD*lq肋IqNv&ΰ_*Nz[:ms|eP3ݗ>3׋-sœUoyi,UFAsRC0GjV6=I{j! /4g;o&?)GB|)ylyǸnNaU ( gF!h f'=BnPjuiK/Ěսpc8$s8=싩:++'QBo≭#8V*pjT٬_)Zv鈉G$@xGUp<Bu@#Ѻ`꩹7R0$.T\ZqkrBzC\ 1*anjkawPAyEH>dJ%[eCba@p(3` Rp֜p$Z $rB&m@0eqP(εye' f4i `$d&Uԃ*Ӑ7B)fBAqm $:ʉ"b[h`# lrY G2 gMs"ڂ{EspaC*Sj$A6O^T#m#bD )hv 03rŒ@ ٝ@ wuڸKV ȯg7^O n [B ֟gw0+Dw邼yw T\޾1#woո<|7Ss5XMpˇ`I̅wO[΂,v濿odp5bF>U O$TW_=M7 +aO 4*L - C֖W5,`,V"2C@S-śZdq!S#}srIAܱ@!cUqDG@%a`vy"tN#}:[zHCL9rۯ:y䢙 4/)Z_:l@GZwky)]znO t,o*w! $#8Ub#)z;q$ꈼ*Em,{T Tʕ$ӧlݣ}uO!9E5aק^rhIS'9=S{?>jZ=:jf={ԴF'GM!ӵ:7lT'sogtp7ύUu:̛̚Np>| G&ʼnd x'O;8O;ph|mST*F㮷&^M6W!BҫZz_jhjh:k?-YSg壭k#GD]L1؈b(NrL{y󑍤RNUOVQ4ly~|uF951Xe,u`BI AaPvLcMW ӱOEvSP<_pϠʛuƔ(Ls?S,S0¯O -'~F%$#5&W(Lka?*eC,3,PfNA+$K){1 )yrşzg;^Du{"$+ 'c-(+$gP`<c&2IB,HgwH `2=e!O",B]J)$OHta0U]C39^~[AĔ%ZFk1{O }9v bqT*0L*1Bn1sت^(|Ϋndj sjn)<,XL)6!{Zo r+)T;M`R'd3#D9G &rϋ(:&wVxBdZh /!cxz0,}d0,G7эLFgݛnE]շ9EYY Wh,qxwSIڈ$\i~w^*P$_{tXj$q0ߖui1P b0'FH4zE (STRJEuEo:1d_}8Zs@? M҆f;^+ފCw&kZM3/k;'%K 8*?cq ¹GKf/a^]EQ|#fa50X|¿0W<,n>-07yø/ĘX8hqow=hvCN ܓrG0~V$}!i[SD{O&_:~iQ)YdĂ(­JO,8㬒DPiZTɴo"KQp8rDLl3(ch^VTO9lyJӹ(.ü|vRʆ]H=-2aw gs>_ ݐ ˶!H4uŽ1A h58ScT CDܸ3Hm9Rk|'7jV1 (LR(6^GrGhJOL %8CԊˎI\7}iUf"/HՊbbVU+j`Щۻtr_נ6pyx9g/e,̈́QD+̇2ȅd?y?qf˖31R f>-;!V7^{VJ4~@˪, 9c_"'ˁ OxMFz[XbaC6Ơ=EŎw3Km TN%A}$ ?HWF/!o ?!$/uL"9XHN_8ըBI24ŀDd=H5Adsr6˜ )I̩pO-O̰Ü2xM/j0`ce )aHɴB@v Ӌ-h鹑d׮Kզ35D/-IW#y7>"m@YhI'([-sA5smڎ~a!@s9/6VIs.ոb,t-᪑&/<9I 0|>JX͘*G-Ԣpm bd΄ ̟A&i˵1 v&MNi$ƩSPv?* KJE*#*H\Zx "d -\Lܝ=q֨k1f"IE>H> yƘ J`V8rV۱X\z<#u=d=z,AG'>d&nQd4/ك@WsFmC  S eBlyMݵn{:t?@z=#&1UzZqmP96I!TS F14tFW jE<B2F#KIHޢ'Z3ϸOz~ rJa+0qR;N8L8qwrܼҶ[^\ƶq%'+詒ģxDQɖu6DOl;Di-NĞ#5=' o -s u. ξ&:]m A3{)x9Yju\?9O5RidԆ؅$<Jo${%{%{%{T l|g"v*7YH2DJ*: .FYlFē$E, l˝=5B_"5_ݗOis  4Q9d@s$i.S$Yѽ[HiHM/&cu&"҂A J$"H$ʛƎ'R2jhp>Q&PeB"nERDRΞ]8 P&*:rz1u}aAhebǨGNE3K LL{]\xDF8y/6RT yV{g!Ii50~#a/iםZ, ;%Qk$^'o IՂqQP#nAKi&oCrQi@70GN=!R$"OJDaM0t0ш>IĐd ,E[ZPh!" 0jQY彊?aaojG*r(wbB\YL SjDQbM' E0;rBt:JjB$bȵT}ZУ`?b[V;$|K@q# (!%G M$&!2JFEv$17QB7BӌYBBB!8rBkoߴ2ẠaĭHIXhH$wX,8⫁>᫸9 z. Wi [ Wn9`񝅏bU fmMOb}rz}} #3ߪ-),Gvς4t'7) F&6G!C2$s1qR^Bd V|rl @aYUʕPD (]^w*Ì~'",@USfs8z!eBjLS1 "d9m7A?AB+SoF"7 Q6z(,S'  2 jj:'~| ,rj983gvq<)iR5kKH@fЊ8E3{6^ zU|`<߽fZ!0X S{k[QMZnnș-65>EI?7鷖pAvȅϲ ;vnP1 "~(E((g R(_g`EOm 괔AihQ%||CM c$=ep9g:M}tt fV;%P2L!j)Nkֵ*V 1LUynWMdfl ZJ~|q0.[iɌjKtY4 Y>+ZnF~=rLv#!\D[T-=T/rSQBmۇvkKi&mAigڭ9cvBB޹)Q6{[O[.ݹ{btK(i].f "YU; {# O;Ƕ):#dmyI."#9/6N21#2),cABJbjI >IqKi.:IW`2L1PPG^bp$3t ܅ E*-A-Q5SA-1>Yc^[|]R.HP_lB-=uK?n)LJ)J`8)}M%$G-LI)Ќ3'նGKH2@S GJ- bկ;6K) f 9Jp¨qV4!~]Z0T% z/nݠTMXnqbٻP%&|#ְ2A5cϚqm՟gg$Kz|]! 6A \q±4q{"DZl~}Js6Ǯ ˈu, a5:ۈژԂ{;'6bxٍ`#YHH9ؕG|d-ey+o&+}-Α(وQjЕd vAJަZ#yK-Gզ0K1zlmj1"O;var/home/core/zuul-output/logs/kubelet.log0000644000000000000000001675213315156324411017707 0ustar rootrootMar 17 18:47:35 crc systemd[1]: Starting Kubernetes Kubelet... Mar 17 18:47:36 crc kubenswrapper[5110]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:47:36 crc kubenswrapper[5110]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 17 18:47:36 crc kubenswrapper[5110]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:47:36 crc kubenswrapper[5110]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:47:36 crc kubenswrapper[5110]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 17 18:47:36 crc kubenswrapper[5110]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.449202 5110 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455499 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455545 5110 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455557 5110 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455567 5110 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455577 5110 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455587 5110 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455597 5110 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455610 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455635 5110 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455645 5110 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455653 5110 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455662 5110 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455671 5110 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455679 5110 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455688 5110 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455698 5110 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455706 5110 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455713 5110 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455721 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455731 5110 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455740 5110 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455749 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455758 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455767 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455777 5110 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455786 5110 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455795 5110 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455804 5110 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455812 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455821 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455830 5110 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455840 5110 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455849 5110 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455860 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455870 5110 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455880 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455891 5110 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455901 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455911 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455921 5110 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455931 5110 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455941 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455949 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455959 5110 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455968 5110 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455978 5110 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455988 5110 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.455998 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456007 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456016 5110 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456029 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456039 5110 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456048 5110 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456094 5110 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456103 5110 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456112 5110 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456123 5110 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456132 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456141 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456149 5110 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456158 5110 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456168 5110 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456176 5110 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456186 5110 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456195 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456205 5110 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456217 5110 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456229 5110 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456238 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456251 5110 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456263 5110 feature_gate.go:328] unrecognized feature gate: Example Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456272 5110 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456281 5110 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456288 5110 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456296 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456308 5110 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456318 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456326 5110 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456333 5110 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456341 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456349 5110 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456356 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456364 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456404 5110 feature_gate.go:328] unrecognized feature gate: Example2 Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456411 5110 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.456419 5110 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457411 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457433 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457443 5110 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457452 5110 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457461 5110 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457471 5110 feature_gate.go:328] unrecognized feature gate: Example Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457481 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457489 5110 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457498 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457508 5110 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457517 5110 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457526 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457538 5110 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457553 5110 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457567 5110 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457576 5110 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457585 5110 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457594 5110 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457603 5110 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457612 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457621 5110 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457630 5110 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457641 5110 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457650 5110 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457659 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457668 5110 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457677 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457687 5110 feature_gate.go:328] unrecognized feature gate: Example2 Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457695 5110 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457705 5110 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457714 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457723 5110 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457732 5110 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457742 5110 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457751 5110 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457760 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457769 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457779 5110 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457792 5110 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457804 5110 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457814 5110 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457825 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457838 5110 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457848 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457857 5110 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457889 5110 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457899 5110 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457908 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457918 5110 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457927 5110 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457936 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457946 5110 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457955 5110 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457965 5110 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457974 5110 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457983 5110 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.457992 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458001 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458011 5110 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458020 5110 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458030 5110 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458040 5110 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458050 5110 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458095 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458104 5110 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458113 5110 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458122 5110 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458131 5110 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458140 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458150 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458159 5110 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458177 5110 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458187 5110 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458196 5110 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458205 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458214 5110 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458223 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458232 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458245 5110 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458256 5110 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458266 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458276 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458285 5110 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458294 5110 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458308 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.458317 5110 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459447 5110 flags.go:64] FLAG: --address="0.0.0.0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459489 5110 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459507 5110 flags.go:64] FLAG: --anonymous-auth="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459520 5110 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459542 5110 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459554 5110 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459567 5110 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459582 5110 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459592 5110 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459603 5110 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459614 5110 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459625 5110 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459636 5110 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459646 5110 flags.go:64] FLAG: --cgroup-root="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459657 5110 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459667 5110 flags.go:64] FLAG: --client-ca-file="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459678 5110 flags.go:64] FLAG: --cloud-config="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459689 5110 flags.go:64] FLAG: --cloud-provider="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459699 5110 flags.go:64] FLAG: --cluster-dns="[]" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459711 5110 flags.go:64] FLAG: --cluster-domain="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459722 5110 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459733 5110 flags.go:64] FLAG: --config-dir="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459744 5110 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459756 5110 flags.go:64] FLAG: --container-log-max-files="5" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459770 5110 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459785 5110 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459797 5110 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459810 5110 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459821 5110 flags.go:64] FLAG: --contention-profiling="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459832 5110 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459843 5110 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459854 5110 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459865 5110 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459878 5110 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459889 5110 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459900 5110 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459910 5110 flags.go:64] FLAG: --enable-load-reader="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459921 5110 flags.go:64] FLAG: --enable-server="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459931 5110 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459947 5110 flags.go:64] FLAG: --event-burst="100" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459958 5110 flags.go:64] FLAG: --event-qps="50" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459969 5110 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459980 5110 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.459991 5110 flags.go:64] FLAG: --eviction-hard="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460004 5110 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460015 5110 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460026 5110 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460037 5110 flags.go:64] FLAG: --eviction-soft="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460047 5110 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460103 5110 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460115 5110 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460125 5110 flags.go:64] FLAG: --experimental-mounter-path="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460137 5110 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460148 5110 flags.go:64] FLAG: --fail-swap-on="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460158 5110 flags.go:64] FLAG: --feature-gates="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460176 5110 flags.go:64] FLAG: --file-check-frequency="20s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460188 5110 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460200 5110 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460214 5110 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460224 5110 flags.go:64] FLAG: --healthz-port="10248" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460235 5110 flags.go:64] FLAG: --help="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460246 5110 flags.go:64] FLAG: --hostname-override="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460257 5110 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460268 5110 flags.go:64] FLAG: --http-check-frequency="20s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460279 5110 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460289 5110 flags.go:64] FLAG: --image-credential-provider-config="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460299 5110 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460310 5110 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460321 5110 flags.go:64] FLAG: --image-service-endpoint="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460332 5110 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460343 5110 flags.go:64] FLAG: --kube-api-burst="100" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460353 5110 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460398 5110 flags.go:64] FLAG: --kube-api-qps="50" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460409 5110 flags.go:64] FLAG: --kube-reserved="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460420 5110 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460430 5110 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460442 5110 flags.go:64] FLAG: --kubelet-cgroups="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460451 5110 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460463 5110 flags.go:64] FLAG: --lock-file="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460473 5110 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460483 5110 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460494 5110 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460512 5110 flags.go:64] FLAG: --log-json-split-stream="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460523 5110 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460534 5110 flags.go:64] FLAG: --log-text-split-stream="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460544 5110 flags.go:64] FLAG: --logging-format="text" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460554 5110 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460567 5110 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460578 5110 flags.go:64] FLAG: --manifest-url="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460588 5110 flags.go:64] FLAG: --manifest-url-header="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460603 5110 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460617 5110 flags.go:64] FLAG: --max-open-files="1000000" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460630 5110 flags.go:64] FLAG: --max-pods="110" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460641 5110 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460653 5110 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460663 5110 flags.go:64] FLAG: --memory-manager-policy="None" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460673 5110 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460684 5110 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460694 5110 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460705 5110 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhel" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460734 5110 flags.go:64] FLAG: --node-status-max-images="50" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460745 5110 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460755 5110 flags.go:64] FLAG: --oom-score-adj="-999" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460765 5110 flags.go:64] FLAG: --pod-cidr="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460776 5110 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cc2b30e70040205c2536d01ae5c850be1ed2d775cf13249e50328e5085777977" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460793 5110 flags.go:64] FLAG: --pod-manifest-path="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460803 5110 flags.go:64] FLAG: --pod-max-pids="-1" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460814 5110 flags.go:64] FLAG: --pods-per-core="0" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460824 5110 flags.go:64] FLAG: --port="10250" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460835 5110 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460847 5110 flags.go:64] FLAG: --provider-id="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460858 5110 flags.go:64] FLAG: --qos-reserved="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460868 5110 flags.go:64] FLAG: --read-only-port="10255" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460879 5110 flags.go:64] FLAG: --register-node="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460890 5110 flags.go:64] FLAG: --register-schedulable="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460899 5110 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460934 5110 flags.go:64] FLAG: --registry-burst="10" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460944 5110 flags.go:64] FLAG: --registry-qps="5" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460954 5110 flags.go:64] FLAG: --reserved-cpus="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460964 5110 flags.go:64] FLAG: --reserved-memory="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460976 5110 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460987 5110 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.460997 5110 flags.go:64] FLAG: --rotate-certificates="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461007 5110 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461020 5110 flags.go:64] FLAG: --runonce="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461031 5110 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461042 5110 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461092 5110 flags.go:64] FLAG: --seccomp-default="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461104 5110 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461115 5110 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461126 5110 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461137 5110 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461147 5110 flags.go:64] FLAG: --storage-driver-password="root" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461158 5110 flags.go:64] FLAG: --storage-driver-secure="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461168 5110 flags.go:64] FLAG: --storage-driver-table="stats" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461178 5110 flags.go:64] FLAG: --storage-driver-user="root" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461188 5110 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461199 5110 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461210 5110 flags.go:64] FLAG: --system-cgroups="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461220 5110 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461242 5110 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461252 5110 flags.go:64] FLAG: --tls-cert-file="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461261 5110 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461274 5110 flags.go:64] FLAG: --tls-min-version="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461284 5110 flags.go:64] FLAG: --tls-private-key-file="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461295 5110 flags.go:64] FLAG: --topology-manager-policy="none" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461305 5110 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461316 5110 flags.go:64] FLAG: --topology-manager-scope="container" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461327 5110 flags.go:64] FLAG: --v="2" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461342 5110 flags.go:64] FLAG: --version="false" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461356 5110 flags.go:64] FLAG: --vmodule="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461370 5110 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.461383 5110 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461692 5110 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461712 5110 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461722 5110 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461733 5110 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461747 5110 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461757 5110 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461766 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461775 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461785 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461794 5110 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461804 5110 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461813 5110 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461824 5110 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461833 5110 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461843 5110 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461853 5110 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461862 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461879 5110 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461889 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461901 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461910 5110 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461920 5110 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461929 5110 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461937 5110 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461947 5110 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461957 5110 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461967 5110 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461974 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461981 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461989 5110 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.461996 5110 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462006 5110 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462018 5110 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462027 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462041 5110 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462050 5110 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462098 5110 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462110 5110 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462120 5110 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462129 5110 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462138 5110 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462147 5110 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462156 5110 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462167 5110 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462176 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462185 5110 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462194 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462204 5110 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462212 5110 feature_gate.go:328] unrecognized feature gate: Example2 Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462223 5110 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462233 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462242 5110 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462251 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462260 5110 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462270 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462279 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462288 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462297 5110 feature_gate.go:328] unrecognized feature gate: Example Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462306 5110 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462315 5110 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462325 5110 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462334 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462343 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462352 5110 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462361 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462370 5110 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462380 5110 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462390 5110 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462401 5110 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462410 5110 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462433 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462445 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462454 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462463 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462472 5110 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462481 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462491 5110 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462500 5110 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462510 5110 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462519 5110 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462528 5110 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462537 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462546 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462560 5110 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462571 5110 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.462581 5110 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.462598 5110 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.474262 5110 server.go:530] "Kubelet version" kubeletVersion="v1.33.5" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.474299 5110 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474372 5110 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474380 5110 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474384 5110 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474388 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474392 5110 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474396 5110 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474399 5110 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474403 5110 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474406 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474410 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474413 5110 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474416 5110 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474420 5110 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474423 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474427 5110 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474431 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474434 5110 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474437 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474441 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474444 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474447 5110 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474451 5110 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474454 5110 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474457 5110 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474461 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474464 5110 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474467 5110 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474470 5110 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474474 5110 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474477 5110 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474480 5110 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474485 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474488 5110 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474493 5110 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474497 5110 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474501 5110 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474504 5110 feature_gate.go:328] unrecognized feature gate: Example2 Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474509 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474513 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474517 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474521 5110 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474524 5110 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474528 5110 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474532 5110 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474535 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474541 5110 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474546 5110 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474550 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474554 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474557 5110 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474561 5110 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474565 5110 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474568 5110 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474571 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474575 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474578 5110 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474581 5110 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474584 5110 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474587 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474590 5110 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474593 5110 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474597 5110 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474600 5110 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474603 5110 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474607 5110 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474610 5110 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474614 5110 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474617 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474620 5110 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474624 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474627 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474630 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474633 5110 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474637 5110 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474640 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474643 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474648 5110 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474653 5110 feature_gate.go:328] unrecognized feature gate: Example Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474656 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474660 5110 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474663 5110 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474666 5110 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474670 5110 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474673 5110 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474676 5110 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474679 5110 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.474686 5110 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474812 5110 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474820 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474825 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474829 5110 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474832 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474836 5110 feature_gate.go:328] unrecognized feature gate: Example Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474839 5110 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474842 5110 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474845 5110 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474849 5110 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474853 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474856 5110 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474860 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474863 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474866 5110 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474869 5110 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474873 5110 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474876 5110 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474879 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474882 5110 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474885 5110 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474889 5110 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474892 5110 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474895 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474898 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474901 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474905 5110 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474911 5110 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474914 5110 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474917 5110 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474920 5110 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474924 5110 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474928 5110 feature_gate.go:328] unrecognized feature gate: Example2 Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474931 5110 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474935 5110 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474938 5110 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474941 5110 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474944 5110 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474948 5110 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474951 5110 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474954 5110 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474957 5110 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474960 5110 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474966 5110 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474971 5110 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474976 5110 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474981 5110 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474985 5110 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474989 5110 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474992 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474996 5110 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.474999 5110 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475003 5110 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475006 5110 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475009 5110 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475013 5110 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475016 5110 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475020 5110 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475023 5110 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475029 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475032 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475036 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475039 5110 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475043 5110 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475046 5110 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475050 5110 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475067 5110 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475071 5110 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475075 5110 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475078 5110 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475081 5110 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475085 5110 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475089 5110 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475092 5110 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475096 5110 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475104 5110 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475109 5110 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475113 5110 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475117 5110 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475121 5110 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475124 5110 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475128 5110 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475132 5110 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475136 5110 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475140 5110 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 17 18:47:36 crc kubenswrapper[5110]: W0317 18:47:36.475144 5110 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.475150 5110 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.476002 5110 server.go:962] "Client rotation is on, will bootstrap in background" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.480172 5110 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2025-12-03 08:27:53 +0000 UTC" logger="UnhandledError" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.483546 5110 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.483659 5110 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.484652 5110 server.go:1019] "Starting client certificate rotation" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.484829 5110 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.484910 5110 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.518880 5110 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.523389 5110 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.524585 5110 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.539808 5110 log.go:25] "Validated CRI v1 runtime API" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.590329 5110 log.go:25] "Validated CRI v1 image API" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.595965 5110 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.601442 5110 fs.go:135] Filesystem UUIDs: map[19e76f87-96b8-4794-9744-0b33dca22d5b:/dev/vda3 2026-03-17-18-41-14-00:/dev/sr0 5eb7c122-420e-4494-80ec-41664070d7b6:/dev/vda4 7B77-95E7:/dev/vda2] Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.601478 5110 fs.go:136] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:45 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:46 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.617694 5110 manager.go:217] Machine: {Timestamp:2026-03-17 18:47:36.614781685 +0000 UTC m=+0.647393227 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33649926144 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:80bc4fba336e4ca1bc9d28a8be52a356 SystemUUID:32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4 BootID:d232fc5f-58b2-431a-8078-c05e8a241121 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16824963072 Type:vfs Inodes:4107657 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6729986048 Type:vfs Inodes:819200 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6545408 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16824963072 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:45 Capacity:3364990976 Type:vfs Inodes:821531 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:46 Capacity:1073741824 Type:vfs Inodes:4107657 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8e:f9:a4 Speed:0 Mtu:1500} {Name:br-int MacAddress:b2:a9:9f:57:07:84 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8e:f9:a4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:a6:c3:9f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c6:25:3a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:52:a7:ce Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b0:56:3b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3a:7d:a5:f6:8e:60 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7a:3f:f3:63:95:3c Speed:0 Mtu:1500} {Name:tap0 MacAddress:5a:94:ef:e4:0c:ee Speed:10 Mtu:1500}] Topology:[{Id:0 Memory:33649926144 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.617937 5110 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.618100 5110 manager.go:233] Version: {KernelVersion:5.14.0-570.57.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20251021-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.620278 5110 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.620317 5110 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.620535 5110 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.620544 5110 container_manager_linux.go:306] "Creating device plugin manager" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.620565 5110 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.621491 5110 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.622239 5110 state_mem.go:36] "Initialized new in-memory state store" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.622383 5110 server.go:1267] "Using root directory" path="/var/lib/kubelet" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.625571 5110 kubelet.go:491] "Attempting to sync node with API server" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.625599 5110 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.625623 5110 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.625639 5110 kubelet.go:397] "Adding apiserver pod source" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.625658 5110 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.630293 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.630411 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.630990 5110 state_checkpoint.go:81] "State checkpoint: restored pod resource state from checkpoint" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.631023 5110 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.632901 5110 state_checkpoint.go:81] "State checkpoint: restored pod resource state from checkpoint" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.632926 5110 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.637621 5110 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.5-3.rhaos4.20.gitd0ea985.el9" apiVersion="v1" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.637861 5110 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-server-current.pem" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.638569 5110 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639540 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639569 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639581 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639591 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639600 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639610 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639620 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639629 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639639 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639654 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.639667 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.640130 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.641504 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.641526 5110 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.642030 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.663238 5110 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.663288 5110 server.go:1295] "Started kubelet" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.663567 5110 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.663582 5110 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.663793 5110 server_v1.go:47] "podresources" method="list" useActivePods=true Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.664446 5110 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 18:47:36 crc systemd[1]: Started Kubernetes Kubelet. Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.665295 5110 server.go:317] "Adding debug handlers to kubelet server" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.667239 5110 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.667621 5110 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.668278 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="200ms" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.668554 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.668575 5110 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.668601 5110 volume_manager.go:295] "The desired_state_of_world populator starts" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.673080 5110 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.673172 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.668513 5110 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189db5610b805182 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.663257474 +0000 UTC m=+0.695868996,LastTimestamp:2026-03-17 18:47:36.663257474 +0000 UTC m=+0.695868996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.678484 5110 factory.go:55] Registering systemd factory Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.678563 5110 factory.go:223] Registration of the systemd container factory successfully Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.678880 5110 factory.go:153] Registering CRI-O factory Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.678901 5110 factory.go:223] Registration of the crio container factory successfully Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.678970 5110 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.679009 5110 factory.go:103] Registering Raw factory Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.679027 5110 manager.go:1196] Started watching for new ooms in manager Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.679479 5110 manager.go:319] Starting recovery of all containers Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.717748 5110 manager.go:324] Recovery completed Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.734588 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.737725 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.738206 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.738220 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.740833 5110 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.741459 5110 cpu_manager.go:222] "Starting CPU manager" policy="none" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.741480 5110 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.741496 5110 state_mem.go:36] "Initialized new in-memory state store" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.743483 5110 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.743522 5110 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.743551 5110 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.743563 5110 kubelet.go:2451] "Starting kubelet main sync loop" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.743684 5110 kubelet.go:2475] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744118 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744201 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744216 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744231 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744243 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7e2c886-118e-43bb-bef1-c78134de392b" volumeName="kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744254 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744266 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744278 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" volumeName="kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744294 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744305 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744317 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744329 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744342 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744354 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744370 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744381 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744394 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744405 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744418 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744429 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744440 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744453 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" volumeName="kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744465 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744477 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" volumeName="kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744489 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744500 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744512 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744526 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744542 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" volumeName="kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744553 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744565 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744577 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="42a11a02-47e1-488f-b270-2679d3298b0e" volumeName="kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744589 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744601 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744612 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744627 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744639 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744651 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744664 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a14caf222afb62aaabdc47808b6f944" volumeName="kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744677 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744690 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744704 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744715 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5ebfebf6-3ecd-458e-943f-bb25b52e2718" volumeName="kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744728 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744742 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744755 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744769 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744782 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744795 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744807 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744820 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e093be35-bb62-4843-b2e8-094545761610" volumeName="kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744833 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" volumeName="kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744845 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744858 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744870 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5ebfebf6-3ecd-458e-943f-bb25b52e2718" volumeName="kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744883 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744901 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744913 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744925 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744937 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744949 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744962 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744975 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.744989 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" volumeName="kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745000 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a208c9c2-333b-4b4a-be0d-bc32ec38a821" volumeName="kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745013 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" volumeName="kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745025 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745036 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745048 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745087 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745102 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745113 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745125 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745136 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745147 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745158 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745171 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745183 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745195 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745207 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745220 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745232 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" volumeName="kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745243 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745254 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745266 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745278 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="34177974-8d82-49d2-a763-391d0df3bbd8" volumeName="kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745291 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745302 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745315 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a208c9c2-333b-4b4a-be0d-bc32ec38a821" volumeName="kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745326 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745339 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f863fff9-286a-45fa-b8f0-8a86994b8440" volumeName="kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745352 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745363 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745377 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745389 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745401 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745413 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745425 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745460 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745472 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745483 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745496 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745517 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af41de71-79cf-4590-bbe9-9e8b848862cb" volumeName="kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745530 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745541 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745553 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745565 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745576 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745588 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745599 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745620 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745631 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745654 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745666 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a14caf222afb62aaabdc47808b6f944" volumeName="kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745677 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745688 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745700 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745712 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745723 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745735 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745747 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745759 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="17b87002-b798-480a-8e17-83053d698239" volumeName="kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745770 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="42a11a02-47e1-488f-b270-2679d3298b0e" volumeName="kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745783 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745794 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745806 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745819 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745831 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745843 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745854 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745865 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745878 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745889 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7e2c886-118e-43bb-bef1-c78134de392b" volumeName="kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745902 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745914 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745924 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745935 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745945 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745956 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745968 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745982 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.745993 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746006 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746017 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746029 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746041 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746078 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20c5c5b4bed930554494851fe3cb2b2a" volumeName="kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746097 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746112 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746127 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746149 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746163 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746175 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746186 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746199 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746211 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746222 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746233 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746246 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746258 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746273 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746284 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746295 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746307 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746319 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746331 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ee8fbd3-1f81-4666-96da-5afc70819f1a" volumeName="kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746343 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746354 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746366 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746379 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746421 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746434 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746445 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ee8fbd3-1f81-4666-96da-5afc70819f1a" volumeName="kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746458 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746472 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746485 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746497 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746509 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746520 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746532 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746544 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" volumeName="kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746555 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746569 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746581 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746592 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746604 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746616 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746632 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746645 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.746658 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.747341 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.749834 5110 policy_none.go:49] "None policy: Start" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.749849 5110 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.749860 5110 state_mem.go:35] "Initializing new in-memory state store" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750511 5110 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1264ac67579ad07e7e9003054d44fe40dd55285a4b2f7dc74e48be1aee0868a/globalmount" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750544 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750560 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750573 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750585 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750599 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750611 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750624 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750635 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750648 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750661 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b638b8f4bb0070e40528db779baf6a2" volumeName="kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750672 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0effdbcf-dd7d-404d-9d48-77536d665a5d" volumeName="kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750687 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750698 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750711 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750723 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750736 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750748 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750759 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750771 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750782 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750795 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750808 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750821 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750835 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750846 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750858 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750870 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750882 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="34177974-8d82-49d2-a763-391d0df3bbd8" volumeName="kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750897 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750909 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750921 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750933 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750944 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750956 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.750996 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751008 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751020 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751030 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751041 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751075 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751095 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751108 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751120 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751132 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751144 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" volumeName="kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751157 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751168 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751182 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751193 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751204 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751216 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751227 5110 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm" seLinuxMountContext="" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751238 5110 reconstruct.go:97] "Volume reconstruction finished" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.751247 5110 reconciler.go:26] "Reconciler: start to sync state" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.773685 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.788111 5110 manager.go:341] "Starting Device Plugin manager" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.788347 5110 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.788366 5110 server.go:85] "Starting device plugin registration server" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.788766 5110 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.788783 5110 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.788993 5110 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.789134 5110 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.789144 5110 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.796733 5110 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.796777 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.844684 5110 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.844864 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.845516 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.845544 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.845553 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846146 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846245 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846293 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846486 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846511 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846521 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846754 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846804 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846816 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.846989 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847128 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847187 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847470 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847540 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847560 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847625 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847653 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.847668 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.849722 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.850285 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.850352 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.852727 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.852752 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.852761 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.852913 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.852924 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.852951 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.856757 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.856885 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.856928 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.857938 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.857983 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.858003 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.858319 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.858345 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.858356 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.859105 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.859155 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.859681 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.859714 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.859724 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.869329 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="400ms" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.889611 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.890307 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.890348 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.890359 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.890382 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.890829 5110 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.900357 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.925171 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.946856 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953225 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953803 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953871 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953900 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-run-kubernetes\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953946 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953973 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.953996 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954017 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954038 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954102 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954316 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954368 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954380 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954377 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954501 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954554 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954581 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954612 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954639 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954670 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954695 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954716 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954742 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954773 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.954797 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.955273 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.955287 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.955399 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-run-kubernetes\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.955781 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: I0317 18:47:36.955832 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.970035 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:36 crc kubenswrapper[5110]: E0317 18:47:36.976015 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056414 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056488 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056523 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056554 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056564 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056643 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056582 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056692 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056710 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056744 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056749 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056777 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056798 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056814 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056842 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056847 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056880 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056913 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056932 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056946 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056973 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056980 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057000 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057028 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.056886 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057101 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057117 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057148 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057029 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057162 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057227 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.057268 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.091129 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.092539 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.092635 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.092687 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.092724 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:47:37 crc kubenswrapper[5110]: E0317 18:47:37.093611 5110 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.201676 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.226188 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.248226 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: W0317 18:47:37.251377 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e08c320b1e9e2405e6e0107bdf7eeb4.slice/crio-20818b330818dc91fcf467bab67a1c3972a5115f03d19356064d284cd8c433a6 WatchSource:0}: Error finding container 20818b330818dc91fcf467bab67a1c3972a5115f03d19356064d284cd8c433a6: Status 404 returned error can't find the container with id 20818b330818dc91fcf467bab67a1c3972a5115f03d19356064d284cd8c433a6 Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.261668 5110 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 18:47:37 crc kubenswrapper[5110]: E0317 18:47:37.270441 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="800ms" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.270582 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.276727 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:37 crc kubenswrapper[5110]: W0317 18:47:37.277614 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20c5c5b4bed930554494851fe3cb2b2a.slice/crio-c0ab80d3888654098fb9175d1c20f61d39ca10ec97a4a3e958d2ede856a51cdf WatchSource:0}: Error finding container c0ab80d3888654098fb9175d1c20f61d39ca10ec97a4a3e958d2ede856a51cdf: Status 404 returned error can't find the container with id c0ab80d3888654098fb9175d1c20f61d39ca10ec97a4a3e958d2ede856a51cdf Mar 17 18:47:37 crc kubenswrapper[5110]: W0317 18:47:37.283787 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a14caf222afb62aaabdc47808b6f944.slice/crio-6940844bfae70b8faec37783ef0f62a7fde79e938fc43932f3b00ab286a5dad5 WatchSource:0}: Error finding container 6940844bfae70b8faec37783ef0f62a7fde79e938fc43932f3b00ab286a5dad5: Status 404 returned error can't find the container with id 6940844bfae70b8faec37783ef0f62a7fde79e938fc43932f3b00ab286a5dad5 Mar 17 18:47:37 crc kubenswrapper[5110]: W0317 18:47:37.307775 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f0bc7fcb0822a2c13eb2d22cd8c0641.slice/crio-001c29075fef9426f46d6bfc795aa3f4edebe49fa5969dea3c2552287e2eade4 WatchSource:0}: Error finding container 001c29075fef9426f46d6bfc795aa3f4edebe49fa5969dea3c2552287e2eade4: Status 404 returned error can't find the container with id 001c29075fef9426f46d6bfc795aa3f4edebe49fa5969dea3c2552287e2eade4 Mar 17 18:47:37 crc kubenswrapper[5110]: W0317 18:47:37.310282 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b638b8f4bb0070e40528db779baf6a2.slice/crio-605f1a7cb887086a14d09985c872529060dbaea88387cf719566c907a9024bce WatchSource:0}: Error finding container 605f1a7cb887086a14d09985c872529060dbaea88387cf719566c907a9024bce: Status 404 returned error can't find the container with id 605f1a7cb887086a14d09985c872529060dbaea88387cf719566c907a9024bce Mar 17 18:47:37 crc kubenswrapper[5110]: E0317 18:47:37.491799 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.494130 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.495225 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.495285 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.495298 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.495328 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:47:37 crc kubenswrapper[5110]: E0317 18:47:37.495776 5110 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.643161 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.749019 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"605f1a7cb887086a14d09985c872529060dbaea88387cf719566c907a9024bce"} Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.750235 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"001c29075fef9426f46d6bfc795aa3f4edebe49fa5969dea3c2552287e2eade4"} Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.751270 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"6940844bfae70b8faec37783ef0f62a7fde79e938fc43932f3b00ab286a5dad5"} Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.752207 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"c0ab80d3888654098fb9175d1c20f61d39ca10ec97a4a3e958d2ede856a51cdf"} Mar 17 18:47:37 crc kubenswrapper[5110]: I0317 18:47:37.753081 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerStarted","Data":"20818b330818dc91fcf467bab67a1c3972a5115f03d19356064d284cd8c433a6"} Mar 17 18:47:37 crc kubenswrapper[5110]: E0317 18:47:37.905234 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 17 18:47:37 crc kubenswrapper[5110]: E0317 18:47:37.997717 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.071930 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="1.6s" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.262634 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.296796 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.298096 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.298178 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.298199 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.298231 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.298792 5110 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.620832 5110 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.622657 5110 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.643402 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.757654 5110 generic.go:358] "Generic (PLEG): container finished" podID="4e08c320b1e9e2405e6e0107bdf7eeb4" containerID="13c01ebb653a6bdaad86850e3714878a54746b6c7cc793fd32d45dade13e7e85" exitCode=0 Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.757727 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerDied","Data":"13c01ebb653a6bdaad86850e3714878a54746b6c7cc793fd32d45dade13e7e85"} Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.757758 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.758281 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.758306 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.758314 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.758447 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.761015 5110 generic.go:358] "Generic (PLEG): container finished" podID="0b638b8f4bb0070e40528db779baf6a2" containerID="49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237" exitCode=0 Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.761109 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerDied","Data":"49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237"} Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.761373 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.762312 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.762336 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.762345 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.762492 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.763606 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"892b0c9c14fabad6da32657ff71c3a8b34d373f96898fd63f4c538fc11e3552a"} Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.763651 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3"} Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.765110 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520" exitCode=0 Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.765164 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520"} Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.765586 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.766659 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.766680 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.766689 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.766834 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767192 5110 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="a29edae24756ec647faeaacc3417fe2dc3d9edbd4dd08de37d3d930ea9060a00" exitCode=0 Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767239 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"a29edae24756ec647faeaacc3417fe2dc3d9edbd4dd08de37d3d930ea9060a00"} Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767351 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767892 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767949 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767980 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.767992 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:38 crc kubenswrapper[5110]: E0317 18:47:38.768384 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.770084 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.770157 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:38 crc kubenswrapper[5110]: I0317 18:47:38.770185 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.588174 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.643584 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.230:6443: connect: connection refused Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.673335 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="3.2s" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.777202 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"53b7011b6f8c381cac7dfbb307f74d4a1bc3c33189e417c20d240acaf720cec6"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.777238 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"f91a3deb02a02a3162c98c11c3614f0692f7ccc8366e5de75ac9d079142ac507"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.777248 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"e19e709ef87fb9cb61061fe5d588545d4e6145cb814255fd180b74cae14a991f"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.777279 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.779707 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.779733 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.779741 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.779901 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.782272 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"8686316b4620286368e9ae51f21b41341ab76b0a6173743fbb9345e41cb8cfc2"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.782296 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"deff5a49b7611cfa357b185827a99ebb293931b96fcd7303036b7a26b5b350c5"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.782355 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.782992 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.783019 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.783030 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.783418 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.785401 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.785801 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.785826 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.785838 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.792163 5110 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="652b9629cee1d27b3cd8bb6186c5d479f4614b794bd8b255c10e38cc41d4ac0a" exitCode=0 Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.792191 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"652b9629cee1d27b3cd8bb6186c5d479f4614b794bd8b255c10e38cc41d4ac0a"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.792304 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.797443 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.797467 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.797476 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.797630 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.799505 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerStarted","Data":"8545c60fc2c493fe31c8c9c36d9eac4f122e0feeea1c08863a4e28302fa86510"} Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.799570 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.803711 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.803730 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.803739 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.803857 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.899813 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.900834 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.900887 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.900902 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:39 crc kubenswrapper[5110]: I0317 18:47:39.900938 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:47:39 crc kubenswrapper[5110]: E0317 18:47:39.901484 5110 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.230:6443: connect: connection refused" node="crc" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.044942 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.054559 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.126976 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.230:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.804402 5110 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="65cf9a08bb479461bfc81b04b4184f9881e3354555c9a3b0c77409d162e9b15c" exitCode=0 Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.804521 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"65cf9a08bb479461bfc81b04b4184f9881e3354555c9a3b0c77409d162e9b15c"} Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.804947 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.805925 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.805980 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.805999 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.806315 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.809084 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"e7e240c309d6b71a597d5ec3f64bffdb103813e71b8290eadd85c4f214db451a"} Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.809178 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.809280 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.809307 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.809348 5110 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.809405 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810367 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810411 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810430 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810408 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810478 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810501 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810548 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810578 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810592 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810593 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810637 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:40 crc kubenswrapper[5110]: I0317 18:47:40.810648 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.810928 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.810922 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.811403 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:40 crc kubenswrapper[5110]: E0317 18:47:40.811605 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.177992 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.685124 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.694614 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.812730 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"18527c5fa347f446246a519571a862dc05435325ff8bb4de538c7834fa60f67c"} Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.812774 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"6f8f146b9f6264d1b969799f87c68b356d3d701c5836b314a29278bd223e63ee"} Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.812788 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"bfcb8305c913d0503cc502dcbfbdedb73b129c8d9b557cf913169eda6efd6d1f"} Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.812852 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.812922 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.812999 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.813498 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.813541 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.813559 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:41 crc kubenswrapper[5110]: E0317 18:47:41.813902 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.815340 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.815368 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:41 crc kubenswrapper[5110]: I0317 18:47:41.815379 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:41 crc kubenswrapper[5110]: E0317 18:47:41.815679 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.433188 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.652775 5110 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.821480 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"e30ed6978ad51bf9888576bae1beb1a9c545283b784e4b0aa1d996642a002616"} Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.821539 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"638c38a4a2517c3d40ecaeafeb3aff8c788d7679eec37d03bbf83b1d3227cdce"} Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.821700 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.821719 5110 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.821777 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.821835 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.822805 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.822828 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.822864 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.822888 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.822990 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.823015 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.822870 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.823104 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:42 crc kubenswrapper[5110]: I0317 18:47:42.823129 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:42 crc kubenswrapper[5110]: E0317 18:47:42.823385 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:42 crc kubenswrapper[5110]: E0317 18:47:42.823555 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:42 crc kubenswrapper[5110]: E0317 18:47:42.823838 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.045306 5110 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.045450 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.101948 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.103304 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.103389 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.103415 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.103457 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.824693 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.824898 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.825737 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.825810 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.825836 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.825930 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.825973 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:43 crc kubenswrapper[5110]: I0317 18:47:43.825992 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:43 crc kubenswrapper[5110]: E0317 18:47:43.826543 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:43 crc kubenswrapper[5110]: E0317 18:47:43.826818 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.040305 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.040558 5110 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.040617 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.041736 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.041803 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.041828 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:45 crc kubenswrapper[5110]: E0317 18:47:45.042379 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.717168 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.717458 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.718669 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.718758 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:45 crc kubenswrapper[5110]: I0317 18:47:45.718793 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:45 crc kubenswrapper[5110]: E0317 18:47:45.719700 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:46 crc kubenswrapper[5110]: E0317 18:47:46.797034 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:47:46 crc kubenswrapper[5110]: I0317 18:47:46.904422 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:46 crc kubenswrapper[5110]: I0317 18:47:46.904723 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:46 crc kubenswrapper[5110]: I0317 18:47:46.906389 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:46 crc kubenswrapper[5110]: I0317 18:47:46.906429 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:46 crc kubenswrapper[5110]: I0317 18:47:46.906442 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:46 crc kubenswrapper[5110]: E0317 18:47:46.906735 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:47 crc kubenswrapper[5110]: I0317 18:47:47.786928 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-etcd/etcd-crc" Mar 17 18:47:47 crc kubenswrapper[5110]: I0317 18:47:47.787303 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:47 crc kubenswrapper[5110]: I0317 18:47:47.788325 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:47 crc kubenswrapper[5110]: I0317 18:47:47.788384 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:47 crc kubenswrapper[5110]: I0317 18:47:47.788403 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:47 crc kubenswrapper[5110]: E0317 18:47:47.789119 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:48 crc kubenswrapper[5110]: I0317 18:47:48.762316 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 17 18:47:48 crc kubenswrapper[5110]: I0317 18:47:48.762649 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:48 crc kubenswrapper[5110]: I0317 18:47:48.763788 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:48 crc kubenswrapper[5110]: I0317 18:47:48.763865 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:48 crc kubenswrapper[5110]: I0317 18:47:48.763886 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:48 crc kubenswrapper[5110]: E0317 18:47:48.764719 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:50 crc kubenswrapper[5110]: I0317 18:47:50.643490 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 17 18:47:51 crc kubenswrapper[5110]: I0317 18:47:51.256364 5110 trace.go:236] Trace[1817426674]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Mar-2026 18:47:41.254) (total time: 10001ms): Mar 17 18:47:51 crc kubenswrapper[5110]: Trace[1817426674]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:47:51.256) Mar 17 18:47:51 crc kubenswrapper[5110]: Trace[1817426674]: [10.00148466s] [10.00148466s] END Mar 17 18:47:51 crc kubenswrapper[5110]: E0317 18:47:51.256402 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 17 18:47:51 crc kubenswrapper[5110]: E0317 18:47:51.658814 5110 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.189db5610b805182 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.663257474 +0000 UTC m=+0.695868996,LastTimestamp:2026-03-17 18:47:36.663257474 +0000 UTC m=+0.695868996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:47:51 crc kubenswrapper[5110]: I0317 18:47:51.840307 5110 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 18:47:51 crc kubenswrapper[5110]: I0317 18:47:51.840374 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 17 18:47:51 crc kubenswrapper[5110]: I0317 18:47:51.849846 5110 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 18:47:51 crc kubenswrapper[5110]: I0317 18:47:51.849952 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 17 18:47:52 crc kubenswrapper[5110]: I0317 18:47:52.444199 5110 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]log ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]etcd ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-api-request-count-filter ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-startkubeinformers ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/generic-apiserver-start-informers ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/priority-and-fairness-config-consumer ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/priority-and-fairness-filter ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-apiextensions-informers ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-apiextensions-controllers ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/crd-informer-synced ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-system-namespaces-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-cluster-authentication-info-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-legacy-token-tracking-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-service-ip-repair-controllers ok Mar 17 18:47:52 crc kubenswrapper[5110]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Mar 17 18:47:52 crc kubenswrapper[5110]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/priority-and-fairness-config-producer ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/bootstrap-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-kubernetes-service-cidr-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/start-kube-aggregator-informers ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-status-local-available-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-status-remote-available-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-registration-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-wait-for-first-sync ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-discovery-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/kube-apiserver-autoregistration ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]autoregister-completion ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-openapi-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: [+]poststarthook/apiservice-openapiv3-controller ok Mar 17 18:47:52 crc kubenswrapper[5110]: livez check failed Mar 17 18:47:52 crc kubenswrapper[5110]: I0317 18:47:52.444302 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:47:52 crc kubenswrapper[5110]: E0317 18:47:52.875278 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Mar 17 18:47:53 crc kubenswrapper[5110]: I0317 18:47:53.046455 5110 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 18:47:53 crc kubenswrapper[5110]: I0317 18:47:53.046549 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 18:47:54 crc kubenswrapper[5110]: E0317 18:47:54.783713 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 17 18:47:56 crc kubenswrapper[5110]: E0317 18:47:56.797350 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.848743 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.849602 5110 trace.go:236] Trace[1190968981]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Mar-2026 18:47:46.017) (total time: 10831ms): Mar 17 18:47:56 crc kubenswrapper[5110]: Trace[1190968981]: ---"Objects listed" error:services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope 10831ms (18:47:56.849) Mar 17 18:47:56 crc kubenswrapper[5110]: Trace[1190968981]: [10.831941392s] [10.831941392s] END Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.849656 5110 trace.go:236] Trace[768847733]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Mar-2026 18:47:43.651) (total time: 13198ms): Mar 17 18:47:56 crc kubenswrapper[5110]: Trace[768847733]: ---"Objects listed" error:runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope 13198ms (18:47:56.849) Mar 17 18:47:56 crc kubenswrapper[5110]: Trace[768847733]: [13.19832029s] [13.19832029s] END Mar 17 18:47:56 crc kubenswrapper[5110]: E0317 18:47:56.849713 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 17 18:47:56 crc kubenswrapper[5110]: E0317 18:47:56.849662 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.849736 5110 trace.go:236] Trace[491740986]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Mar-2026 18:47:43.553) (total time: 13296ms): Mar 17 18:47:56 crc kubenswrapper[5110]: Trace[491740986]: ---"Objects listed" error:nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope 13296ms (18:47:56.849) Mar 17 18:47:56 crc kubenswrapper[5110]: Trace[491740986]: [13.296417362s] [13.296417362s] END Mar 17 18:47:56 crc kubenswrapper[5110]: E0317 18:47:56.849763 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 17 18:47:56 crc kubenswrapper[5110]: E0317 18:47:56.857122 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.867591 5110 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.917972 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.918256 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.919973 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.920019 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.920033 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:56 crc kubenswrapper[5110]: E0317 18:47:56.920430 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.947377 5110 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44154->192.168.126.11:17697: read: connection reset by peer" start-of-body= Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.947396 5110 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50428->192.168.126.11:17697: read: connection reset by peer" start-of-body= Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.947442 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44154->192.168.126.11:17697: read: connection reset by peer" Mar 17 18:47:56 crc kubenswrapper[5110]: I0317 18:47:56.947514 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50428->192.168.126.11:17697: read: connection reset by peer" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.439237 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.439484 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.439799 5110 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.439950 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.440409 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.440461 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.440476 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:57 crc kubenswrapper[5110]: E0317 18:47:57.440933 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.443913 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.647388 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.864917 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.867387 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="e7e240c309d6b71a597d5ec3f64bffdb103813e71b8290eadd85c4f214db451a" exitCode=255 Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.867475 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"e7e240c309d6b71a597d5ec3f64bffdb103813e71b8290eadd85c4f214db451a"} Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.867602 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.868323 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.868415 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.868436 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:57 crc kubenswrapper[5110]: E0317 18:47:57.868906 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:57 crc kubenswrapper[5110]: I0317 18:47:57.869358 5110 scope.go:117] "RemoveContainer" containerID="e7e240c309d6b71a597d5ec3f64bffdb103813e71b8290eadd85c4f214db451a" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.647347 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.795949 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.796289 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.797216 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.797332 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.797416 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:58 crc kubenswrapper[5110]: E0317 18:47:58.797913 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.810695 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.871663 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.873452 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3"} Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.873653 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.873680 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.874396 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.874441 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.874459 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.874400 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.874529 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:58 crc kubenswrapper[5110]: I0317 18:47:58.874541 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:58 crc kubenswrapper[5110]: E0317 18:47:58.874820 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:58 crc kubenswrapper[5110]: E0317 18:47:58.875467 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:59 crc kubenswrapper[5110]: E0317 18:47:59.284547 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.646818 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.878880 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.879942 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.882499 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" exitCode=255 Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.882580 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3"} Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.882637 5110 scope.go:117] "RemoveContainer" containerID="e7e240c309d6b71a597d5ec3f64bffdb103813e71b8290eadd85c4f214db451a" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.882846 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.883624 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.883677 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.883696 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:47:59 crc kubenswrapper[5110]: E0317 18:47:59.884297 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:47:59 crc kubenswrapper[5110]: I0317 18:47:59.884655 5110 scope.go:117] "RemoveContainer" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" Mar 17 18:47:59 crc kubenswrapper[5110]: E0317 18:47:59.884940 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.053123 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.053388 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.054405 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.054505 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.054526 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:00 crc kubenswrapper[5110]: E0317 18:48:00.054959 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.061151 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.652140 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.889637 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.892560 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.892787 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.893435 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.893467 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.893471 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.893506 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.893519 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.893477 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:00 crc kubenswrapper[5110]: E0317 18:48:00.893887 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:00 crc kubenswrapper[5110]: E0317 18:48:00.894255 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:00 crc kubenswrapper[5110]: I0317 18:48:00.894596 5110 scope.go:117] "RemoveContainer" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" Mar 17 18:48:00 crc kubenswrapper[5110]: E0317 18:48:00.894859 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.647257 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.664835 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610b805182 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.663257474 +0000 UTC m=+0.695868996,LastTimestamp:2026-03-17 18:47:36.663257474 +0000 UTC m=+0.695868996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.669644 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.678428 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.684421 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.690300 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5611336691f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.792631583 +0000 UTC m=+0.825243105,LastTimestamp:2026-03-17 18:47:36.792631583 +0000 UTC m=+0.825243105,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.696247 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.845531878 +0000 UTC m=+0.878143400,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.701620 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.845548409 +0000 UTC m=+0.878159921,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.716014 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff83de6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.845556989 +0000 UTC m=+0.878168511,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.722521 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.846499154 +0000 UTC m=+0.879110676,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.733777 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.846516784 +0000 UTC m=+0.879128306,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.739476 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff83de6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.846524974 +0000 UTC m=+0.879136496,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.744721 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.846779431 +0000 UTC m=+0.879390953,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.750584 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.846811492 +0000 UTC m=+0.879423014,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.755981 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff83de6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.846820592 +0000 UTC m=+0.879432114,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.762811 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.84749177 +0000 UTC m=+0.880103302,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.769726 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.847551281 +0000 UTC m=+0.880162813,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.776195 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff83de6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.847568672 +0000 UTC m=+0.880180214,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.783467 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.847642674 +0000 UTC m=+0.880254216,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.790284 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.847661924 +0000 UTC m=+0.880273466,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.794590 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff83de6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.847676025 +0000 UTC m=+0.880287567,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.800173 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.852745857 +0000 UTC m=+0.885357379,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.805229 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.852756678 +0000 UTC m=+0.885368200,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.810758 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff83de6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff83de6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738225638 +0000 UTC m=+0.770837160,LastTimestamp:2026-03-17 18:47:36.852765608 +0000 UTC m=+0.885377130,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.816404 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff7ac8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff7ac8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738188427 +0000 UTC m=+0.770799969,LastTimestamp:2026-03-17 18:47:36.852920022 +0000 UTC m=+0.885531544,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.822703 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189db5610ff80d76\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189db5610ff80d76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:36.738213238 +0000 UTC m=+0.770824760,LastTimestamp:2026-03-17 18:47:36.852946073 +0000 UTC m=+0.885557595,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.830278 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189db5612f35473b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.262319419 +0000 UTC m=+1.294930951,LastTimestamp:2026-03-17 18:47:37.262319419 +0000 UTC m=+1.294930951,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.835735 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561306d098d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.282750861 +0000 UTC m=+1.315362383,LastTimestamp:2026-03-17 18:47:37.282750861 +0000 UTC m=+1.315362383,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.841045 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db56130e3dca3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.290538147 +0000 UTC m=+1.323149739,LastTimestamp:2026-03-17 18:47:37.290538147 +0000 UTC m=+1.323149739,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.846599 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db561322088af openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.311291567 +0000 UTC m=+1.343903089,LastTimestamp:2026-03-17 18:47:37.311291567 +0000 UTC m=+1.343903089,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.850439 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db56132509ade openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.31444195 +0000 UTC m=+1.347053472,LastTimestamp:2026-03-17 18:47:37.31444195 +0000 UTC m=+1.347053472,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.854478 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db5615778a88e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container: wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.937823886 +0000 UTC m=+1.970435408,LastTimestamp:2026-03-17 18:47:37.937823886 +0000 UTC m=+1.970435408,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.858690 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db561577d45cc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container: kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.938126284 +0000 UTC m=+1.970737806,LastTimestamp:2026-03-17 18:47:37.938126284 +0000 UTC m=+1.970737806,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.863713 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561577dcdef openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.938161135 +0000 UTC m=+1.970772677,LastTimestamp:2026-03-17 18:47:37.938161135 +0000 UTC m=+1.970772677,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.868776 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189db561577f7395 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.938269077 +0000 UTC m=+1.970880599,LastTimestamp:2026-03-17 18:47:37.938269077 +0000 UTC m=+1.970880599,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.874207 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db5615786620b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.938723339 +0000 UTC m=+1.971334861,LastTimestamp:2026-03-17 18:47:37.938723339 +0000 UTC m=+1.971334861,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.879261 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5615828402d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.949331501 +0000 UTC m=+1.981943023,LastTimestamp:2026-03-17 18:47:37.949331501 +0000 UTC m=+1.981943023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.883119 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5615833e4f5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.950094581 +0000 UTC m=+1.982706103,LastTimestamp:2026-03-17 18:47:37.950094581 +0000 UTC m=+1.982706103,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.887658 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56158661922 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.953384738 +0000 UTC m=+1.985996260,LastTimestamp:2026-03-17 18:47:37.953384738 +0000 UTC m=+1.985996260,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.889433 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561588e4e60 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.956019808 +0000 UTC m=+1.988631330,LastTimestamp:2026-03-17 18:47:37.956019808 +0000 UTC m=+1.988631330,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.893260 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189db561588eb51c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.956046108 +0000 UTC m=+1.988657630,LastTimestamp:2026-03-17 18:47:37.956046108 +0000 UTC m=+1.988657630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.898191 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db5615890cb76 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:37.956182902 +0000 UTC m=+1.988794424,LastTimestamp:2026-03-17 18:47:37.956182902 +0000 UTC m=+1.988794424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.902385 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5616b346427 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container: cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.268894247 +0000 UTC m=+2.301505809,LastTimestamp:2026-03-17 18:47:38.268894247 +0000 UTC m=+2.301505809,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.906363 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5616c026429 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.282394665 +0000 UTC m=+2.315006227,LastTimestamp:2026-03-17 18:47:38.282394665 +0000 UTC m=+2.315006227,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.910031 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5616c14fc3f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.283613247 +0000 UTC m=+2.316224809,LastTimestamp:2026-03-17 18:47:38.283613247 +0000 UTC m=+2.316224809,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.915258 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189db5618874bda0 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.75965072 +0000 UTC m=+2.792262242,LastTimestamp:2026-03-17 18:47:38.75965072 +0000 UTC m=+2.792262242,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.919613 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db56188ca2b6c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.765249388 +0000 UTC m=+2.797860910,LastTimestamp:2026-03-17 18:47:38.765249388 +0000 UTC m=+2.797860910,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.924795 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db56188f0b059 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.767773785 +0000 UTC m=+2.800385307,LastTimestamp:2026-03-17 18:47:38.767773785 +0000 UTC m=+2.800385307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.929294 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561891f0eaa openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.770812586 +0000 UTC m=+2.803424118,LastTimestamp:2026-03-17 18:47:38.770812586 +0000 UTC m=+2.803424118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.934026 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5618a49d0a6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container: kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.790391974 +0000 UTC m=+2.823003496,LastTimestamp:2026-03-17 18:47:38.790391974 +0000 UTC m=+2.823003496,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.939250 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5618b94f61b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.812093979 +0000 UTC m=+2.844705501,LastTimestamp:2026-03-17 18:47:38.812093979 +0000 UTC m=+2.844705501,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.944362 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db5618baa32f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:38.813485816 +0000 UTC m=+2.846097348,LastTimestamp:2026-03-17 18:47:38.813485816 +0000 UTC m=+2.846097348,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.948537 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db56197826a2d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container: kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.012205101 +0000 UTC m=+3.044816623,LastTimestamp:2026-03-17 18:47:39.012205101 +0000 UTC m=+3.044816623,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.949128 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.949421 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.950321 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.950372 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.950387 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.950908 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:01 crc kubenswrapper[5110]: I0317 18:48:01.951239 5110 scope.go:117] "RemoveContainer" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.951486 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.953881 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189db56197a9f6f0 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container: kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.01479704 +0000 UTC m=+3.047408572,LastTimestamp:2026-03-17 18:47:39.01479704 +0000 UTC m=+3.047408572,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.958347 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db56197aa3714 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container: kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.01481346 +0000 UTC m=+3.047424982,LastTimestamp:2026-03-17 18:47:39.01481346 +0000 UTC m=+3.047424982,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.964105 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56197ab8b03 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container: etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.014900483 +0000 UTC m=+3.047512005,LastTimestamp:2026-03-17 18:47:39.014900483 +0000 UTC m=+3.047512005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.969358 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db56197b85b7d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container: kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.015740285 +0000 UTC m=+3.048351807,LastTimestamp:2026-03-17 18:47:39.015740285 +0000 UTC m=+3.048351807,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.974730 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561982938c1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.023136961 +0000 UTC m=+3.055748483,LastTimestamp:2026-03-17 18:47:39.023136961 +0000 UTC m=+3.055748483,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.979412 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561983c8191 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.024400785 +0000 UTC m=+3.057012307,LastTimestamp:2026-03-17 18:47:39.024400785 +0000 UTC m=+3.057012307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.983148 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189db56198619845 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.026831429 +0000 UTC m=+3.059442951,LastTimestamp:2026-03-17 18:47:39.026831429 +0000 UTC m=+3.059442951,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.989015 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db561991792d1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.038757585 +0000 UTC m=+3.071369117,LastTimestamp:2026-03-17 18:47:39.038757585 +0000 UTC m=+3.071369117,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.993202 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db5619923aa7c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.039550076 +0000 UTC m=+3.072161598,LastTimestamp:2026-03-17 18:47:39.039550076 +0000 UTC m=+3.072161598,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:01 crc kubenswrapper[5110]: E0317 18:48:01.997309 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db5619930ea37 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.040418359 +0000 UTC m=+3.073029881,LastTimestamp:2026-03-17 18:47:39.040418359 +0000 UTC m=+3.073029881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.001089 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561998256d4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.04575458 +0000 UTC m=+3.078366102,LastTimestamp:2026-03-17 18:47:39.04575458 +0000 UTC m=+3.078366102,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.005151 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561a2d745d8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container: kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.202315736 +0000 UTC m=+3.234927258,LastTimestamp:2026-03-17 18:47:39.202315736 +0000 UTC m=+3.234927258,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.009614 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561a3c12bd9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container: kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.217644505 +0000 UTC m=+3.250256027,LastTimestamp:2026-03-17 18:47:39.217644505 +0000 UTC m=+3.250256027,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.013973 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561a3db33d1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.219350481 +0000 UTC m=+3.251962003,LastTimestamp:2026-03-17 18:47:39.219350481 +0000 UTC m=+3.251962003,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.018449 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561a3e98d3c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.220290876 +0000 UTC m=+3.252902398,LastTimestamp:2026-03-17 18:47:39.220290876 +0000 UTC m=+3.252902398,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.024890 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561a4e94086 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.237048454 +0000 UTC m=+3.269659976,LastTimestamp:2026-03-17 18:47:39.237048454 +0000 UTC m=+3.269659976,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.030374 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561a4f9ec19 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.238140953 +0000 UTC m=+3.270752475,LastTimestamp:2026-03-17 18:47:39.238140953 +0000 UTC m=+3.270752475,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.035181 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561ada98edc openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container: kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.383869148 +0000 UTC m=+3.416480670,LastTimestamp:2026-03-17 18:47:39.383869148 +0000 UTC m=+3.416480670,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.040502 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189db561aec18388 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.402216328 +0000 UTC m=+3.434827850,LastTimestamp:2026-03-17 18:47:39.402216328 +0000 UTC m=+3.434827850,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.044719 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561af44c720 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container: kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.410818848 +0000 UTC m=+3.443430370,LastTimestamp:2026-03-17 18:47:39.410818848 +0000 UTC m=+3.443430370,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.049462 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561affa98fa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.422734586 +0000 UTC m=+3.455346108,LastTimestamp:2026-03-17 18:47:39.422734586 +0000 UTC m=+3.455346108,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.053271 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561b009e4e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.423737063 +0000 UTC m=+3.456348585,LastTimestamp:2026-03-17 18:47:39.423737063 +0000 UTC m=+3.456348585,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.057844 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561bebc74eb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container: kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.670320363 +0000 UTC m=+3.702931885,LastTimestamp:2026-03-17 18:47:39.670320363 +0000 UTC m=+3.702931885,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.062501 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561bf92e422 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.684373538 +0000 UTC m=+3.716985060,LastTimestamp:2026-03-17 18:47:39.684373538 +0000 UTC m=+3.716985060,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.068764 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561bfa64390 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.685643152 +0000 UTC m=+3.718254674,LastTimestamp:2026-03-17 18:47:39.685643152 +0000 UTC m=+3.718254674,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.073731 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561c66dbe10 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.799379472 +0000 UTC m=+3.831990994,LastTimestamp:2026-03-17 18:47:39.799379472 +0000 UTC m=+3.831990994,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.078242 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561cccaeadd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.906149085 +0000 UTC m=+3.938760607,LastTimestamp:2026-03-17 18:47:39.906149085 +0000 UTC m=+3.938760607,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.082316 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561cd836840 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.918239808 +0000 UTC m=+3.950851330,LastTimestamp:2026-03-17 18:47:39.918239808 +0000 UTC m=+3.950851330,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.088873 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561d1d7ee04 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container: etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.99088794 +0000 UTC m=+4.023499462,LastTimestamp:2026-03-17 18:47:39.99088794 +0000 UTC m=+4.023499462,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.096938 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db561d2d2b6e9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:40.007323369 +0000 UTC m=+4.039934901,LastTimestamp:2026-03-17 18:47:40.007323369 +0000 UTC m=+4.039934901,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.101933 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db5620286190a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:40.807608586 +0000 UTC m=+4.840220148,LastTimestamp:2026-03-17 18:47:40.807608586 +0000 UTC m=+4.840220148,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.107608 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56211e95f68 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container: etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.065772904 +0000 UTC m=+5.098384416,LastTimestamp:2026-03-17 18:47:41.065772904 +0000 UTC m=+5.098384416,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.112873 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db562125b7ed6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.073252054 +0000 UTC m=+5.105863566,LastTimestamp:2026-03-17 18:47:41.073252054 +0000 UTC m=+5.105863566,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.118949 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db5621268d845 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.074126917 +0000 UTC m=+5.106738439,LastTimestamp:2026-03-17 18:47:41.074126917 +0000 UTC m=+5.106738439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.123000 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56220914d5b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container: etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.311659355 +0000 UTC m=+5.344270917,LastTimestamp:2026-03-17 18:47:41.311659355 +0000 UTC m=+5.344270917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.127738 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56221ba79ff openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.331134975 +0000 UTC m=+5.363746537,LastTimestamp:2026-03-17 18:47:41.331134975 +0000 UTC m=+5.363746537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.132566 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56221d13bb3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.332626355 +0000 UTC m=+5.365237897,LastTimestamp:2026-03-17 18:47:41.332626355 +0000 UTC m=+5.365237897,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.137702 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db5623183a733 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container: etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.595977523 +0000 UTC m=+5.628589045,LastTimestamp:2026-03-17 18:47:41.595977523 +0000 UTC m=+5.628589045,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.141528 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56232a9cc96 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.615254678 +0000 UTC m=+5.647866230,LastTimestamp:2026-03-17 18:47:41.615254678 +0000 UTC m=+5.647866230,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.146272 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db56232be28c2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.616588994 +0000 UTC m=+5.649200506,LastTimestamp:2026-03-17 18:47:41.616588994 +0000 UTC m=+5.649200506,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.150789 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db562417713bb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container: etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.863588795 +0000 UTC m=+5.896200327,LastTimestamp:2026-03-17 18:47:41.863588795 +0000 UTC m=+5.896200327,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.154467 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db562425e4812 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.87874101 +0000 UTC m=+5.911352542,LastTimestamp:2026-03-17 18:47:41.87874101 +0000 UTC m=+5.911352542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.158283 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db5624281ed6e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:41.881077102 +0000 UTC m=+5.913688634,LastTimestamp:2026-03-17 18:47:41.881077102 +0000 UTC m=+5.913688634,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.162272 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db5624ea5b1b9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container: etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:42.084747705 +0000 UTC m=+6.117359267,LastTimestamp:2026-03-17 18:47:42.084747705 +0000 UTC m=+6.117359267,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.167136 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189db5624fc4a25f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:42.103552607 +0000 UTC m=+6.136164159,LastTimestamp:2026-03-17 18:47:42.103552607 +0000 UTC m=+6.136164159,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.172016 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-controller-manager-crc.189db56287e86370 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://localhost:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 18:48:02 crc kubenswrapper[5110]: body: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:43.045419888 +0000 UTC m=+7.078031410,LastTimestamp:2026-03-17 18:47:43.045419888 +0000 UTC m=+7.078031410,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.180069 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db56287e9a0de openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:43.04550115 +0000 UTC m=+7.078112682,LastTimestamp:2026-03-17 18:47:43.04550115 +0000 UTC m=+7.078112682,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.187616 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-apiserver-crc.189db5649420622c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 17 18:48:02 crc kubenswrapper[5110]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 18:48:02 crc kubenswrapper[5110]: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:51.840350764 +0000 UTC m=+15.872962286,LastTimestamp:2026-03-17 18:47:51.840350764 +0000 UTC m=+15.872962286,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.192730 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db56494210c0f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:51.840394255 +0000 UTC m=+15.873005777,LastTimestamp:2026-03-17 18:47:51.840394255 +0000 UTC m=+15.873005777,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.196554 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db5649420622c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-apiserver-crc.189db5649420622c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 17 18:48:02 crc kubenswrapper[5110]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 18:48:02 crc kubenswrapper[5110]: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:51.840350764 +0000 UTC m=+15.872962286,LastTimestamp:2026-03-17 18:47:51.849916599 +0000 UTC m=+15.882528161,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.201665 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db56494210c0f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db56494210c0f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:51.840394255 +0000 UTC m=+15.873005777,LastTimestamp:2026-03-17 18:47:51.849981161 +0000 UTC m=+15.882592693,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.207988 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-apiserver-crc.189db564b81f6901 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 500 Mar 17 18:48:02 crc kubenswrapper[5110]: body: [+]ping ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]log ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]etcd ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-api-request-count-filter ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-startkubeinformers ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/generic-apiserver-start-informers ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/priority-and-fairness-config-consumer ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/priority-and-fairness-filter ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-apiextensions-informers ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-apiextensions-controllers ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/crd-informer-synced ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-system-namespaces-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-cluster-authentication-info-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-legacy-token-tracking-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-service-ip-repair-controllers ok Mar 17 18:48:02 crc kubenswrapper[5110]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Mar 17 18:48:02 crc kubenswrapper[5110]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/priority-and-fairness-config-producer ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/bootstrap-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-kubernetes-service-cidr-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/start-kube-aggregator-informers ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-status-local-available-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-status-remote-available-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-registration-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-wait-for-first-sync ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-discovery-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/kube-apiserver-autoregistration ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]autoregister-completion ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-openapi-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: [+]poststarthook/apiservice-openapiv3-controller ok Mar 17 18:48:02 crc kubenswrapper[5110]: livez check failed Mar 17 18:48:02 crc kubenswrapper[5110]: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:52.444266753 +0000 UTC m=+16.476878315,LastTimestamp:2026-03-17 18:47:52.444266753 +0000 UTC m=+16.476878315,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.211998 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db564b8208397 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:52.444339095 +0000 UTC m=+16.476950647,LastTimestamp:2026-03-17 18:47:52.444339095 +0000 UTC m=+16.476950647,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.215736 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189db56287e86370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-controller-manager-crc.189db56287e86370 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://localhost:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 18:48:02 crc kubenswrapper[5110]: body: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:43.045419888 +0000 UTC m=+7.078031410,LastTimestamp:2026-03-17 18:47:53.046515877 +0000 UTC m=+17.079127399,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.219709 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189db56287e9a0de\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189db56287e9a0de openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:43.04550115 +0000 UTC m=+7.078112682,LastTimestamp:2026-03-17 18:47:53.046574609 +0000 UTC m=+17.079186131,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.223713 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-apiserver-crc.189db565c4881743 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:44154->192.168.126.11:17697: read: connection reset by peer Mar 17 18:48:02 crc kubenswrapper[5110]: body: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:56.947420995 +0000 UTC m=+20.980032517,LastTimestamp:2026-03-17 18:47:56.947420995 +0000 UTC m=+20.980032517,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.231873 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db565c488b398 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44154->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:56.947461016 +0000 UTC m=+20.980072538,LastTimestamp:2026-03-17 18:47:56.947461016 +0000 UTC m=+20.980072538,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.237559 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-apiserver-crc.189db565c4882503 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:50428->192.168.126.11:17697: read: connection reset by peer Mar 17 18:48:02 crc kubenswrapper[5110]: body: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:56.947424515 +0000 UTC m=+20.980036057,LastTimestamp:2026-03-17 18:47:56.947424515 +0000 UTC m=+20.980036057,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.243106 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db565c489cbb8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50428->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:56.947532728 +0000 UTC m=+20.980144260,LastTimestamp:2026-03-17 18:47:56.947532728 +0000 UTC m=+20.980144260,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.248267 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 18:48:02 crc kubenswrapper[5110]: &Event{ObjectMeta:{kube-apiserver-crc.189db565e1e1f827 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 17 18:48:02 crc kubenswrapper[5110]: body: Mar 17 18:48:02 crc kubenswrapper[5110]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:57.439850535 +0000 UTC m=+21.472462107,LastTimestamp:2026-03-17 18:47:57.439850535 +0000 UTC m=+21.472462107,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 18:48:02 crc kubenswrapper[5110]: > Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.252238 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db565e1e49fd1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:57.440024529 +0000 UTC m=+21.472636061,LastTimestamp:2026-03-17 18:47:57.440024529 +0000 UTC m=+21.472636061,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.257555 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db561bfa64390\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561bfa64390 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.685643152 +0000 UTC m=+3.718254674,LastTimestamp:2026-03-17 18:47:57.871069839 +0000 UTC m=+21.903681371,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.262452 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db561cccaeadd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561cccaeadd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.906149085 +0000 UTC m=+3.938760607,LastTimestamp:2026-03-17 18:47:58.113811536 +0000 UTC m=+22.146423058,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.267743 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db561cd836840\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561cd836840 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.918239808 +0000 UTC m=+3.950851330,LastTimestamp:2026-03-17 18:47:58.126665179 +0000 UTC m=+22.159276701,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.273511 5110 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.279038 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db566739e783f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:48:00.894820625 +0000 UTC m=+24.927432167,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: E0317 18:48:02.283532 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db566739e783f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:48:01.951447012 +0000 UTC m=+25.984058534,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:02 crc kubenswrapper[5110]: I0317 18:48:02.649566 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:03 crc kubenswrapper[5110]: I0317 18:48:03.257882 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:03 crc kubenswrapper[5110]: I0317 18:48:03.258828 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:03 crc kubenswrapper[5110]: I0317 18:48:03.258891 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:03 crc kubenswrapper[5110]: I0317 18:48:03.258919 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:03 crc kubenswrapper[5110]: I0317 18:48:03.258967 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:03 crc kubenswrapper[5110]: E0317 18:48:03.272817 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:48:03 crc kubenswrapper[5110]: I0317 18:48:03.647262 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:03 crc kubenswrapper[5110]: E0317 18:48:03.818522 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 17 18:48:04 crc kubenswrapper[5110]: E0317 18:48:04.048963 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 17 18:48:04 crc kubenswrapper[5110]: I0317 18:48:04.648789 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:05 crc kubenswrapper[5110]: E0317 18:48:05.564938 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 17 18:48:05 crc kubenswrapper[5110]: I0317 18:48:05.648342 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:06 crc kubenswrapper[5110]: E0317 18:48:06.291023 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:48:06 crc kubenswrapper[5110]: I0317 18:48:06.650122 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:06 crc kubenswrapper[5110]: E0317 18:48:06.797653 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:48:07 crc kubenswrapper[5110]: I0317 18:48:07.649685 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.649421 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.874119 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.875123 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.876154 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.876235 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.876254 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:08 crc kubenswrapper[5110]: E0317 18:48:08.876905 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:08 crc kubenswrapper[5110]: I0317 18:48:08.877300 5110 scope.go:117] "RemoveContainer" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" Mar 17 18:48:08 crc kubenswrapper[5110]: E0317 18:48:08.877638 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:08 crc kubenswrapper[5110]: E0317 18:48:08.882632 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db566739e783f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:48:08.877580205 +0000 UTC m=+32.910191767,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:08 crc kubenswrapper[5110]: E0317 18:48:08.973479 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 17 18:48:09 crc kubenswrapper[5110]: I0317 18:48:09.650778 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:10 crc kubenswrapper[5110]: I0317 18:48:10.273693 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:10 crc kubenswrapper[5110]: I0317 18:48:10.274987 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:10 crc kubenswrapper[5110]: I0317 18:48:10.275117 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:10 crc kubenswrapper[5110]: I0317 18:48:10.275152 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:10 crc kubenswrapper[5110]: I0317 18:48:10.275212 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:10 crc kubenswrapper[5110]: E0317 18:48:10.289020 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:48:10 crc kubenswrapper[5110]: I0317 18:48:10.649050 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:11 crc kubenswrapper[5110]: I0317 18:48:11.647911 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:12 crc kubenswrapper[5110]: I0317 18:48:12.648593 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:13 crc kubenswrapper[5110]: E0317 18:48:13.294009 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:48:13 crc kubenswrapper[5110]: I0317 18:48:13.648298 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:14 crc kubenswrapper[5110]: I0317 18:48:14.647297 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:15 crc kubenswrapper[5110]: I0317 18:48:15.649508 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:16 crc kubenswrapper[5110]: I0317 18:48:16.647965 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:16 crc kubenswrapper[5110]: E0317 18:48:16.797957 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:48:17 crc kubenswrapper[5110]: I0317 18:48:17.290135 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:17 crc kubenswrapper[5110]: I0317 18:48:17.291157 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:17 crc kubenswrapper[5110]: I0317 18:48:17.291223 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:17 crc kubenswrapper[5110]: I0317 18:48:17.291249 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:17 crc kubenswrapper[5110]: I0317 18:48:17.291293 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:17 crc kubenswrapper[5110]: E0317 18:48:17.307141 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:48:17 crc kubenswrapper[5110]: I0317 18:48:17.650348 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:18 crc kubenswrapper[5110]: I0317 18:48:18.648418 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:19 crc kubenswrapper[5110]: I0317 18:48:19.647701 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:20 crc kubenswrapper[5110]: E0317 18:48:20.303203 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:48:20 crc kubenswrapper[5110]: I0317 18:48:20.651865 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:21 crc kubenswrapper[5110]: I0317 18:48:21.651601 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:22 crc kubenswrapper[5110]: I0317 18:48:22.650011 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:23 crc kubenswrapper[5110]: I0317 18:48:23.648982 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:24 crc kubenswrapper[5110]: E0317 18:48:24.174891 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.308283 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.309722 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.309824 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.309844 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.309893 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:24 crc kubenswrapper[5110]: E0317 18:48:24.325353 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.648522 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.744808 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.745793 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.745856 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.745883 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:24 crc kubenswrapper[5110]: E0317 18:48:24.746665 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:24 crc kubenswrapper[5110]: I0317 18:48:24.747114 5110 scope.go:117] "RemoveContainer" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" Mar 17 18:48:24 crc kubenswrapper[5110]: E0317 18:48:24.756009 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db561bfa64390\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561bfa64390 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.685643152 +0000 UTC m=+3.718254674,LastTimestamp:2026-03-17 18:48:24.74863359 +0000 UTC m=+48.781245152,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:25 crc kubenswrapper[5110]: E0317 18:48:25.078874 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 17 18:48:25 crc kubenswrapper[5110]: E0317 18:48:25.193838 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db561cccaeadd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561cccaeadd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.906149085 +0000 UTC m=+3.938760607,LastTimestamp:2026-03-17 18:48:25.18530721 +0000 UTC m=+49.217918732,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:25 crc kubenswrapper[5110]: E0317 18:48:25.251649 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db561cd836840\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db561cd836840 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:39.918239808 +0000 UTC m=+3.950851330,LastTimestamp:2026-03-17 18:48:25.244225245 +0000 UTC m=+49.276836777,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.337551 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.339401 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a"} Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.339677 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.340510 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.340570 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.340589 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:25 crc kubenswrapper[5110]: E0317 18:48:25.341162 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.650219 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.724234 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.724597 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.726290 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.726352 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:25 crc kubenswrapper[5110]: I0317 18:48:25.726374 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:25 crc kubenswrapper[5110]: E0317 18:48:25.726930 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.344824 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.345438 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.347913 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a" exitCode=255 Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.347955 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a"} Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.348103 5110 scope.go:117] "RemoveContainer" containerID="91993913df9bdd75952097ab4bc3ece74c32727bad8680881af3b3f16979a4b3" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.348319 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.348977 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.349032 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.349048 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:26 crc kubenswrapper[5110]: E0317 18:48:26.349555 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.349911 5110 scope.go:117] "RemoveContainer" containerID="6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a" Mar 17 18:48:26 crc kubenswrapper[5110]: E0317 18:48:26.350170 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:26 crc kubenswrapper[5110]: E0317 18:48:26.357235 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db566739e783f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:48:26.350125669 +0000 UTC m=+50.382737191,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:26 crc kubenswrapper[5110]: I0317 18:48:26.649317 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:26 crc kubenswrapper[5110]: E0317 18:48:26.798501 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:48:27 crc kubenswrapper[5110]: E0317 18:48:27.310728 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:48:27 crc kubenswrapper[5110]: I0317 18:48:27.352891 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 17 18:48:27 crc kubenswrapper[5110]: I0317 18:48:27.651156 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:27 crc kubenswrapper[5110]: E0317 18:48:27.740528 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 17 18:48:28 crc kubenswrapper[5110]: E0317 18:48:28.551302 5110 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 17 18:48:28 crc kubenswrapper[5110]: I0317 18:48:28.649207 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:29 crc kubenswrapper[5110]: I0317 18:48:29.649958 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:30 crc kubenswrapper[5110]: I0317 18:48:30.648034 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.326276 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.327344 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.327397 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.327407 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.327433 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:31 crc kubenswrapper[5110]: E0317 18:48:31.341849 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.649997 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.949606 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.949906 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.950870 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.950920 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.950934 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:31 crc kubenswrapper[5110]: E0317 18:48:31.951385 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:31 crc kubenswrapper[5110]: I0317 18:48:31.951680 5110 scope.go:117] "RemoveContainer" containerID="6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a" Mar 17 18:48:31 crc kubenswrapper[5110]: E0317 18:48:31.951888 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:31 crc kubenswrapper[5110]: E0317 18:48:31.957274 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db566739e783f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:48:31.951854279 +0000 UTC m=+55.984465801,Count:6,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:32 crc kubenswrapper[5110]: I0317 18:48:32.648211 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:33 crc kubenswrapper[5110]: I0317 18:48:33.650343 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:34 crc kubenswrapper[5110]: E0317 18:48:34.319240 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:48:34 crc kubenswrapper[5110]: I0317 18:48:34.648678 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.340836 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.341100 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.341971 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.342040 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.342087 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:35 crc kubenswrapper[5110]: E0317 18:48:35.342696 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.343376 5110 scope.go:117] "RemoveContainer" containerID="6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a" Mar 17 18:48:35 crc kubenswrapper[5110]: E0317 18:48:35.343703 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:35 crc kubenswrapper[5110]: E0317 18:48:35.348577 5110 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189db566739e783f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189db566739e783f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:47:59.884900415 +0000 UTC m=+23.917511967,LastTimestamp:2026-03-17 18:48:35.343651811 +0000 UTC m=+59.376263363,Count:7,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:48:35 crc kubenswrapper[5110]: I0317 18:48:35.648201 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:36 crc kubenswrapper[5110]: I0317 18:48:36.649934 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:36 crc kubenswrapper[5110]: E0317 18:48:36.799550 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:48:37 crc kubenswrapper[5110]: I0317 18:48:37.647418 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:38 crc kubenswrapper[5110]: I0317 18:48:38.342191 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:38 crc kubenswrapper[5110]: I0317 18:48:38.343112 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:38 crc kubenswrapper[5110]: I0317 18:48:38.343168 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:38 crc kubenswrapper[5110]: I0317 18:48:38.343181 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:38 crc kubenswrapper[5110]: I0317 18:48:38.343204 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:38 crc kubenswrapper[5110]: E0317 18:48:38.351885 5110 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 18:48:38 crc kubenswrapper[5110]: I0317 18:48:38.650034 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:39 crc kubenswrapper[5110]: I0317 18:48:39.649483 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:40 crc kubenswrapper[5110]: I0317 18:48:40.647565 5110 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 18:48:41 crc kubenswrapper[5110]: E0317 18:48:41.323398 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 18:48:41 crc kubenswrapper[5110]: I0317 18:48:41.495629 5110 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-pqm89" Mar 17 18:48:41 crc kubenswrapper[5110]: I0317 18:48:41.502770 5110 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-pqm89" Mar 17 18:48:41 crc kubenswrapper[5110]: I0317 18:48:41.544020 5110 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 17 18:48:42 crc kubenswrapper[5110]: I0317 18:48:42.484907 5110 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 17 18:48:42 crc kubenswrapper[5110]: I0317 18:48:42.504464 5110 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2026-04-16 18:43:41 +0000 UTC" deadline="2026-04-12 20:25:54.182522267 +0000 UTC" Mar 17 18:48:42 crc kubenswrapper[5110]: I0317 18:48:42.504509 5110 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="625h37m11.67802305s" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.351997 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.353361 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.353434 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.353455 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.353634 5110 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.371307 5110 kubelet_node_status.go:127] "Node was previously registered" node="crc" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.371611 5110 kubelet_node_status.go:81] "Successfully registered node" node="crc" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.371647 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.375102 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.375147 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.375159 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.375175 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.375185 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:45Z","lastTransitionTime":"2026-03-17T18:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.394158 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.404334 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.404408 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.404426 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.404449 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.404468 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:45Z","lastTransitionTime":"2026-03-17T18:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.421326 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.431144 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.431229 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.431255 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.431289 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.431310 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:45Z","lastTransitionTime":"2026-03-17T18:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.445112 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.452722 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.452781 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.452799 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.452821 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:45 crc kubenswrapper[5110]: I0317 18:48:45.452838 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:45Z","lastTransitionTime":"2026-03-17T18:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.462302 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.462458 5110 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.462487 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.563124 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.663701 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.764234 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.864698 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:45 crc kubenswrapper[5110]: E0317 18:48:45.965911 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.066890 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.167085 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.268098 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.368328 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.468560 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.568865 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.669360 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.770047 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.800617 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.870546 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:46 crc kubenswrapper[5110]: E0317 18:48:46.970824 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.071698 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.171784 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.272857 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.373708 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.474068 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.574754 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.675317 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: I0317 18:48:47.744103 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:47 crc kubenswrapper[5110]: I0317 18:48:47.745324 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:47 crc kubenswrapper[5110]: I0317 18:48:47.745368 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:47 crc kubenswrapper[5110]: I0317 18:48:47.745380 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.745869 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:47 crc kubenswrapper[5110]: I0317 18:48:47.746141 5110 scope.go:117] "RemoveContainer" containerID="6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.776221 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.876944 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:47 crc kubenswrapper[5110]: E0317 18:48:47.977320 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.078314 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.178955 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.279762 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.379868 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.404995 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.406702 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220"} Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.406913 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.407436 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.407471 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.407483 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.407878 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.480286 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.580714 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.681292 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.744259 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.745330 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.745375 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:48 crc kubenswrapper[5110]: I0317 18:48:48.745386 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.745730 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.782252 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.883176 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:48 crc kubenswrapper[5110]: E0317 18:48:48.984297 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.084944 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.186155 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.287336 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.387631 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.487976 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.588340 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.688761 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.789547 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.890122 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:49 crc kubenswrapper[5110]: E0317 18:48:49.990803 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.091322 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.192086 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.293101 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.393810 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.413667 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.414348 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.416814 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" exitCode=255 Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.416887 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220"} Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.416995 5110 scope.go:117] "RemoveContainer" containerID="6094825e40295b398a84d1b3298431e523d776449b4cd2716922f37f3eadb51a" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.417294 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.418103 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.418141 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.418154 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.418809 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:50 crc kubenswrapper[5110]: I0317 18:48:50.419112 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.419374 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.494939 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.596330 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.697360 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.797705 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.897925 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:50 crc kubenswrapper[5110]: E0317 18:48:50.998986 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.100210 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.201340 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.302347 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.402893 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.421281 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.503355 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.604043 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.704991 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.805494 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.905825 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.949579 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.949812 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.950865 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.950914 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.950931 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.951588 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:51 crc kubenswrapper[5110]: I0317 18:48:51.951961 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:48:51 crc kubenswrapper[5110]: E0317 18:48:51.952294 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.006003 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.107252 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.207925 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.308143 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.408372 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.508549 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.608699 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.708983 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.809331 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:52 crc kubenswrapper[5110]: E0317 18:48:52.909405 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.010412 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.110700 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.211828 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.312183 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.412625 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.512874 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.613869 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.714922 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.815611 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:53 crc kubenswrapper[5110]: E0317 18:48:53.915787 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.016990 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.117828 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.218910 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.319661 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.420251 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.520507 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.620635 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.720978 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.821602 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:54 crc kubenswrapper[5110]: E0317 18:48:54.921750 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.022218 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.123391 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.224348 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.325089 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.426250 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.527002 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.648720 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.748836 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.848981 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.854353 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.859280 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.859327 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.859343 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.859364 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.859379 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:55Z","lastTransitionTime":"2026-03-17T18:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.878098 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.891432 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.891489 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.891507 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.891531 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.891550 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:55Z","lastTransitionTime":"2026-03-17T18:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.903548 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.914631 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.914678 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.914695 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.914719 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.914746 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:55Z","lastTransitionTime":"2026-03-17T18:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.933108 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.944632 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.944717 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.944738 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.944765 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:48:55 crc kubenswrapper[5110]: I0317 18:48:55.944784 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:48:55Z","lastTransitionTime":"2026-03-17T18:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.959978 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.960278 5110 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Mar 17 18:48:55 crc kubenswrapper[5110]: E0317 18:48:55.960315 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.061142 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.161736 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.262327 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.362711 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.463781 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.564600 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.665399 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.766348 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.801217 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.866870 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:56 crc kubenswrapper[5110]: E0317 18:48:56.967975 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.068355 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.169520 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.270560 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.371175 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.472213 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.572903 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.673809 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.774513 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.875042 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:57 crc kubenswrapper[5110]: E0317 18:48:57.976179 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.076977 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.177442 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.278223 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.378773 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: I0317 18:48:58.407538 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:48:58 crc kubenswrapper[5110]: I0317 18:48:58.407826 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:48:58 crc kubenswrapper[5110]: I0317 18:48:58.408939 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:48:58 crc kubenswrapper[5110]: I0317 18:48:58.408989 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:48:58 crc kubenswrapper[5110]: I0317 18:48:58.408998 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.409557 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:48:58 crc kubenswrapper[5110]: I0317 18:48:58.409844 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.410085 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.479419 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.579906 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.681021 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.781876 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.882226 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:58 crc kubenswrapper[5110]: E0317 18:48:58.983328 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.084044 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.184289 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.285395 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.385944 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.486699 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.587262 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.688160 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.788648 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: I0317 18:48:59.812141 5110 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.889373 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:48:59 crc kubenswrapper[5110]: E0317 18:48:59.990024 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.090228 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.190789 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.291236 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.391679 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.492271 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.592718 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.693267 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.794455 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.895126 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:00 crc kubenswrapper[5110]: E0317 18:49:00.996166 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.097113 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.197312 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.297978 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.398830 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.499871 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.600399 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.701577 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.802389 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:01 crc kubenswrapper[5110]: E0317 18:49:01.903141 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.004182 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.104914 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.205944 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.306228 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.406395 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.507261 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.608154 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.708997 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.809762 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:02 crc kubenswrapper[5110]: E0317 18:49:02.910738 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.011157 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.111934 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.212456 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.313437 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.414361 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.514998 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.615811 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.716906 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.817756 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:03 crc kubenswrapper[5110]: E0317 18:49:03.918075 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.019074 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.119241 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.220132 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.321102 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.421309 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: I0317 18:49:04.459024 5110 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.521553 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.622681 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.722886 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.823412 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:04 crc kubenswrapper[5110]: E0317 18:49:04.923778 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.024201 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.124806 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.225212 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.325787 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.426177 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.526757 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.627124 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.727908 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: I0317 18:49:05.744506 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:49:05 crc kubenswrapper[5110]: I0317 18:49:05.745795 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:05 crc kubenswrapper[5110]: I0317 18:49:05.745853 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:05 crc kubenswrapper[5110]: I0317 18:49:05.745871 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.746803 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.828008 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:05 crc kubenswrapper[5110]: E0317 18:49:05.928388 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.029097 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.129557 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.229968 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.285793 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.289879 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.289934 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.289952 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.289975 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.289993 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:06Z","lastTransitionTime":"2026-03-17T18:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.303823 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.308236 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.308294 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.308313 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.308335 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.308353 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:06Z","lastTransitionTime":"2026-03-17T18:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.323749 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.328642 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.328691 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.328702 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.328718 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.328729 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:06Z","lastTransitionTime":"2026-03-17T18:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.338351 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.342835 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.342978 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.343121 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.343252 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:06 crc kubenswrapper[5110]: I0317 18:49:06.343386 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:06Z","lastTransitionTime":"2026-03-17T18:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.355584 5110 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400456Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861256Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d232fc5f-58b2-431a-8078-c05e8a241121\\\",\\\"systemUUID\\\":\\\"32f2ed8d-b82d-4a5e-8804-40a7fd2f20c4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.355718 5110 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.355756 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.456270 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.556602 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.657987 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.758847 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.802434 5110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.859482 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:06 crc kubenswrapper[5110]: E0317 18:49:06.960289 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.061665 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.162217 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.263434 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.364631 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.464742 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.565600 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.666409 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.767166 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.868123 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:07 crc kubenswrapper[5110]: E0317 18:49:07.968954 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.069017 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.169408 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.269510 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.370485 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.471185 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.571690 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.671891 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.772286 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.873357 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:08 crc kubenswrapper[5110]: E0317 18:49:08.974725 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.075018 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.175654 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.276308 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.376916 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.477485 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.577738 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.678532 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: I0317 18:49:09.744442 5110 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 17 18:49:09 crc kubenswrapper[5110]: I0317 18:49:09.745449 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:09 crc kubenswrapper[5110]: I0317 18:49:09.745532 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:09 crc kubenswrapper[5110]: I0317 18:49:09.745558 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.746422 5110 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 17 18:49:09 crc kubenswrapper[5110]: I0317 18:49:09.746896 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.747277 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.778874 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.879904 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:09 crc kubenswrapper[5110]: E0317 18:49:09.980802 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.081795 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.182236 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.282775 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.383526 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.484542 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.585023 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.686155 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.786969 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.887640 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:10 crc kubenswrapper[5110]: E0317 18:49:10.988143 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.088934 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.190142 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.290556 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.391162 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.492173 5110 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.545719 5110 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.568812 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.579318 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.593672 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.593705 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.593714 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.593730 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.593739 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:11Z","lastTransitionTime":"2026-03-17T18:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.674418 5110 apiserver.go:52] "Watching apiserver" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.682136 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.683513 5110 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.684030 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-machine-config-operator/machine-config-daemon-6jbgs","openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5","openshift-network-node-identity/network-node-identity-dgvkt","openshift-network-operator/iptables-alerter-5jnd7","openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488","openshift-dns/node-resolver-6k4l7","openshift-image-registry/node-ca-slqfs","openshift-multus/multus-6kh2q","openshift-multus/multus-additional-cni-plugins-lh6lx","openshift-multus/network-metrics-daemon-m5g58","openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6","openshift-network-diagnostics/network-check-target-fhkjl","openshift-ovn-kubernetes/ovnkube-node-4jphj","openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv"] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.685160 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.686877 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.687172 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.687324 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.687867 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"metrics-tls\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.688258 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.690527 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.690745 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.690650 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.695205 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"env-overrides\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.695421 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-node-identity\"/\"network-node-identity-cert\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699396 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699453 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699471 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699493 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699510 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:11Z","lastTransitionTime":"2026-03-17T18:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699559 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.699829 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.700509 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.702628 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.703583 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"ovnkube-identity-cm\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.709668 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.709788 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.709906 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.712546 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.713292 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.713719 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-tk7bt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.713836 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.717671 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.723785 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.724118 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-tjs74\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.723793 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.724469 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.725144 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.731738 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.731952 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.732017 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-g6kgg\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.733046 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.733486 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.733585 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.734618 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.734906 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.737252 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.740152 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.740418 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-nwglk\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.742853 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.743406 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.745245 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.745539 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.745714 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.746308 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.746568 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.746677 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.746973 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.747447 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.747471 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-l2v2m\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.747610 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.747790 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.748082 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.748306 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.750589 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.750685 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.750686 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.750785 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"proxy-tls\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.752666 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-nl8tp\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.752862 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.752900 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-w9nzh\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.761104 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.770149 5110 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.770460 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.778904 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.779355 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.784461 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.795955 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.802954 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.802993 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.803005 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.803021 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.803033 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:11Z","lastTransitionTime":"2026-03-17T18:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.806894 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-6k4l7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stxn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6k4l7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.821038 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lh6lx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.830392 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7da49224-3b49-41d3-8490-ae2724128e67\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-csf4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-csf4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6jbgs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.838906 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849210 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849244 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849270 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849295 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849315 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849336 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgx6b\" (UniqueName: \"kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b\") pod \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\" (UID: \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849359 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849381 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849401 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849425 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849446 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849466 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849486 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l87hs\" (UniqueName: \"kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs\") pod \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\" (UID: \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849509 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849528 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849549 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849572 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9stx\" (UniqueName: \"kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849596 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849620 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nspp\" (UniqueName: \"kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849668 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849689 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849713 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849734 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849755 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849776 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849834 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849899 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849925 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm9x7\" (UniqueName: \"kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849947 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849970 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w94wk\" (UniqueName: \"kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.849992 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dztfv\" (UniqueName: \"kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850014 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850036 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850074 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850100 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850125 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850146 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850167 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmmzf\" (UniqueName: \"kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850231 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850255 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850304 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850334 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzt4w\" (UniqueName: \"kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w\") pod \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\" (UID: \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850390 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850413 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850438 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850460 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850483 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g8ts\" (UniqueName: \"kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850508 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptkcf\" (UniqueName: \"kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850531 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850556 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850580 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850603 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850625 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850649 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850674 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850696 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850744 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850770 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4tqq\" (UniqueName: \"kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq\") pod \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\" (UID: \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850795 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850816 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850841 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nb9c\" (UniqueName: \"kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850863 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850886 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls\") pod \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\" (UID: \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850910 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsb9b\" (UniqueName: \"kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850933 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850959 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jjkz\" (UniqueName: \"kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851041 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851084 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851112 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851135 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851857 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnxbn\" (UniqueName: \"kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851893 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851922 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851950 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftwb6\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852000 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852024 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tknt7\" (UniqueName: \"kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852048 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852110 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852135 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852159 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852183 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852209 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852233 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir\") pod \"f7e2c886-118e-43bb-bef1-c78134de392b\" (UID: \"f7e2c886-118e-43bb-bef1-c78134de392b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852259 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkdh6\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852505 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852648 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852735 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852779 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852857 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852935 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hckvg\" (UniqueName: \"kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852976 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854263 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26xrl\" (UniqueName: \"kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl\") pod \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\" (UID: \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.850141 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855307 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities" (OuterVolumeSpecName: "utilities") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855410 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq" (OuterVolumeSpecName: "kube-api-access-d4tqq") pod "6ee8fbd3-1f81-4666-96da-5afc70819f1a" (UID: "6ee8fbd3-1f81-4666-96da-5afc70819f1a"). InnerVolumeSpecName "kube-api-access-d4tqq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855428 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855614 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities" (OuterVolumeSpecName: "utilities") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855787 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855813 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856160 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp" (OuterVolumeSpecName: "tmp") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856201 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856335 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqbfk\" (UniqueName: \"kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856351 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856417 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856426 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856432 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856308 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331f5282-47de-4f21-903e-5dfb189879f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T18:48:49Z\\\",\\\"message\\\":\\\" envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"ClientsAllowCBOR\\\\\\\" enabled=false\\\\nW0317 18:48:49.126497 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 18:48:49.126663 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI0317 18:48:49.127547 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-97051896/tls.crt::/tmp/serving-cert-97051896/tls.key\\\\\\\"\\\\nI0317 18:48:49.391789 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 18:48:49.393987 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 18:48:49.394005 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 18:48:49.394028 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 18:48:49.394034 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 18:48:49.397362 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 18:48:49.397406 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0317 18:48:49.397405 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0317 18:48:49.397411 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 18:48:49.397432 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 18:48:49.397438 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 18:48:49.397442 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 18:48:49.397445 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0317 18:48:49.399325 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T18:48:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856475 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856469 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856579 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856617 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856646 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert\") pod \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\" (UID: \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856672 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856695 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856720 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856744 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856766 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856790 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856815 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z4sw\" (UniqueName: \"kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856821 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c" (OuterVolumeSpecName: "kube-api-access-8nb9c") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "kube-api-access-8nb9c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856840 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856866 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856892 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.856895 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857127 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities" (OuterVolumeSpecName: "utilities") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857138 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn" (OuterVolumeSpecName: "kube-api-access-xnxbn") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "kube-api-access-xnxbn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857375 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config" (OuterVolumeSpecName: "console-config") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857544 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "6ee8fbd3-1f81-4666-96da-5afc70819f1a" (UID: "6ee8fbd3-1f81-4666-96da-5afc70819f1a"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857594 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs" (OuterVolumeSpecName: "kube-api-access-l87hs") pod "5ebfebf6-3ecd-458e-943f-bb25b52e2718" (UID: "5ebfebf6-3ecd-458e-943f-bb25b52e2718"). InnerVolumeSpecName "kube-api-access-l87hs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857606 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfzkj\" (UniqueName: \"kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj\") pod \"0effdbcf-dd7d-404d-9d48-77536d665a5d\" (UID: \"0effdbcf-dd7d-404d-9d48-77536d665a5d\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857746 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857805 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857842 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857937 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857984 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4smf\" (UniqueName: \"kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf\") pod \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\" (UID: \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858039 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858111 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858162 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pllx6\" (UniqueName: \"kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858212 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hb7m\" (UniqueName: \"kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858253 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs\") pod \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\" (UID: \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858295 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858332 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twvbl\" (UniqueName: \"kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858475 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858652 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858874 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858927 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858964 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859005 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859044 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859139 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859179 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks6v2\" (UniqueName: \"kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859217 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859254 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859289 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859329 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859371 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859411 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859448 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pskd\" (UniqueName: \"kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859487 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859529 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859567 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859602 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859639 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859680 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859732 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859798 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859868 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5lgh\" (UniqueName: \"kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859993 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860039 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjwtd\" (UniqueName: \"kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860124 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860167 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860216 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxfcv\" (UniqueName: \"kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860256 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860296 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860333 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8zz\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860360 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860387 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7cps\" (UniqueName: \"kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps\") pod \"af41de71-79cf-4590-bbe9-9e8b848862cb\" (UID: \"af41de71-79cf-4590-bbe9-9e8b848862cb\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860419 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4lr\" (UniqueName: \"kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr\") pod \"f7e2c886-118e-43bb-bef1-c78134de392b\" (UID: \"f7e2c886-118e-43bb-bef1-c78134de392b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860446 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj4qr\" (UniqueName: \"kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860478 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860507 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg8nc\" (UniqueName: \"kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860534 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860563 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddlk9\" (UniqueName: \"kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860589 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbc2l\" (UniqueName: \"kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860618 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860644 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860673 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860725 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5rsr\" (UniqueName: \"kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860759 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgrkj\" (UniqueName: \"kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj\") pod \"42a11a02-47e1-488f-b270-2679d3298b0e\" (UID: \"42a11a02-47e1-488f-b270-2679d3298b0e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860786 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860812 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860839 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zth6t\" (UniqueName: \"kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860867 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860893 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860919 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860948 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vsz9\" (UniqueName: \"kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860974 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861169 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861200 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861238 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861275 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861306 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94l9h\" (UniqueName: \"kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861338 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861373 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfp5s\" (UniqueName: \"kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861412 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmqg\" (UniqueName: \"kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861455 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861492 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pddnv\" (UniqueName: \"kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv\") pod \"e093be35-bb62-4843-b2e8-094545761610\" (UID: \"e093be35-bb62-4843-b2e8-094545761610\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861519 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861547 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861573 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861602 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m26jq\" (UniqueName: \"kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861631 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861659 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca\") pod \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\" (UID: \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861702 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls\") pod \"42a11a02-47e1-488f-b270-2679d3298b0e\" (UID: \"42a11a02-47e1-488f-b270-2679d3298b0e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861731 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861762 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861793 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs\") pod \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\" (UID: \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861819 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861848 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861877 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861907 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rmnv\" (UniqueName: \"kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861934 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861963 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861989 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862016 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862042 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862108 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862138 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862166 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862228 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862265 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862293 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862323 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmhf\" (UniqueName: \"kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862353 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zj9\" (UniqueName: \"kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862381 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862410 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lcfw\" (UniqueName: \"kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862440 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert\") pod \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\" (UID: \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862469 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862514 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862557 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grwfz\" (UniqueName: \"kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862604 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862638 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862680 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862748 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862825 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-kubelet\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862859 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-netns\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862895 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-etc-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862920 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-ovn\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862947 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-netns\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862980 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863014 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gfq8\" (UniqueName: \"kubernetes.io/projected/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-kube-api-access-6gfq8\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863040 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-systemd-units\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863095 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-slash\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863122 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863152 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863180 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-multus-certs\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863209 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863240 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863448 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-host\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863479 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-netd\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863511 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8nt2j\" (UniqueName: \"kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863539 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-cnibin\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863564 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-cni-multus\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863611 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863644 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnc69\" (UniqueName: \"kubernetes.io/projected/baee355c-9055-42f1-87c4-3dc7a08b68f6-kube-api-access-fnc69\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863672 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7da49224-3b49-41d3-8490-ae2724128e67-rootfs\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863709 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-config\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863754 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgwk\" (UniqueName: \"kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863803 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-etc-kubernetes\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863850 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-system-cni-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863876 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7da49224-3b49-41d3-8490-ae2724128e67-mcd-auth-proxy-config\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863901 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-serviceca\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863933 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2nhm\" (UniqueName: \"kubernetes.io/projected/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-kube-api-access-d2nhm\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863957 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-var-lib-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863982 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-node-log\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864010 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovn-node-metrics-cert\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864040 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864089 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7346b312-05b5-4475-8d54-82709b69e7ed-cni-binary-copy\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864115 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-k8s-cni-cncf-io\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864145 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864176 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864204 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-conf-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864235 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864267 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m7xz2\" (UniqueName: \"kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864302 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864338 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864365 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-log-socket\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864403 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-hosts-file\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864429 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-system-cni-dir\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864454 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-os-release\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857849 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.857996 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851643 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852909 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.852941 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp" (OuterVolumeSpecName: "tmp") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853118 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6" (OuterVolumeSpecName: "kube-api-access-tkdh6") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "kube-api-access-tkdh6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853130 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853355 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853603 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config" (OuterVolumeSpecName: "config") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.866841 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config" (OuterVolumeSpecName: "config") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853739 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853770 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b" (OuterVolumeSpecName: "kube-api-access-zsb9b") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "kube-api-access-zsb9b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853778 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b" (OuterVolumeSpecName: "kube-api-access-pgx6b") pod "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" (UID: "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4"). InnerVolumeSpecName "kube-api-access-pgx6b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853889 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz" (OuterVolumeSpecName: "kube-api-access-7jjkz") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "kube-api-access-7jjkz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853919 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853828 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca" (OuterVolumeSpecName: "client-ca") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.853989 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854127 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854269 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854259 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854289 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp" (OuterVolumeSpecName: "tmp") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854519 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.854679 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7" (OuterVolumeSpecName: "kube-api-access-tknt7") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "kube-api-access-tknt7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855204 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.855228 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.851447 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp" (OuterVolumeSpecName: "kube-api-access-8nspp") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "kube-api-access-8nspp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858220 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858263 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858408 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config" (OuterVolumeSpecName: "config") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858428 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858594 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6" (OuterVolumeSpecName: "kube-api-access-ftwb6") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "kube-api-access-ftwb6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858615 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config" (OuterVolumeSpecName: "config") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858848 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.858891 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859187 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts" (OuterVolumeSpecName: "kube-api-access-4g8ts") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "kube-api-access-4g8ts". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859329 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859502 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859600 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf" (OuterVolumeSpecName: "kube-api-access-ptkcf") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "kube-api-access-ptkcf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859692 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859745 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities" (OuterVolumeSpecName: "utilities") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859804 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.859899 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860405 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg" (OuterVolumeSpecName: "kube-api-access-hckvg") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "kube-api-access-hckvg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860452 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860646 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj" (OuterVolumeSpecName: "kube-api-access-mfzkj") pod "0effdbcf-dd7d-404d-9d48-77536d665a5d" (UID: "0effdbcf-dd7d-404d-9d48-77536d665a5d"). InnerVolumeSpecName "kube-api-access-mfzkj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.860867 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl" (OuterVolumeSpecName: "kube-api-access-26xrl") pod "a208c9c2-333b-4b4a-be0d-bc32ec38a821" (UID: "a208c9c2-333b-4b4a-be0d-bc32ec38a821"). InnerVolumeSpecName "kube-api-access-26xrl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861196 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config" (OuterVolumeSpecName: "config") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.861815 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx" (OuterVolumeSpecName: "kube-api-access-l9stx") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "kube-api-access-l9stx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862216 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config" (OuterVolumeSpecName: "config") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.862898 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863090 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863612 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.863628 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864130 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864304 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w" (OuterVolumeSpecName: "kube-api-access-rzt4w") pod "a52afe44-fb37-46ed-a1f8-bf39727a3cbe" (UID: "a52afe44-fb37-46ed-a1f8-bf39727a3cbe"). InnerVolumeSpecName "kube-api-access-rzt4w". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864847 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.864864 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf" (OuterVolumeSpecName: "kube-api-access-nmmzf") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "kube-api-access-nmmzf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.865319 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.865470 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.865579 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.865766 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.865955 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config" (OuterVolumeSpecName: "config") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.866177 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.866507 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867224 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.866556 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk" (OuterVolumeSpecName: "kube-api-access-w94wk") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "kube-api-access-w94wk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.866922 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities" (OuterVolumeSpecName: "utilities") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867248 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp" (OuterVolumeSpecName: "tmp") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867702 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-ovn-kubernetes\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867742 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-script-lib\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867781 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867809 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-systemd\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867874 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.867988 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-cni-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868014 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-os-release\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868041 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7346b312-05b5-4475-8d54-82709b69e7ed-multus-daemon-config\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868088 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-bin\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868118 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868147 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868178 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-tmp-dir\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868201 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7da49224-3b49-41d3-8490-ae2724128e67-proxy-tls\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868336 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868362 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868382 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lck5w\" (UniqueName: \"kubernetes.io/projected/7346b312-05b5-4475-8d54-82709b69e7ed-kube-api-access-lck5w\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868412 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868439 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csf4c\" (UniqueName: \"kubernetes.io/projected/7da49224-3b49-41d3-8490-ae2724128e67-kube-api-access-csf4c\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868458 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-cni-bin\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868477 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-kubelet\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868498 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-hostroot\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868521 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stxn5\" (UniqueName: \"kubernetes.io/projected/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-kube-api-access-stxn5\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868612 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cnibin\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868634 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cni-binary-copy\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868653 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b652\" (UniqueName: \"kubernetes.io/projected/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-kube-api-access-8b652\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868670 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868694 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85swk\" (UniqueName: \"kubernetes.io/projected/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-kube-api-access-85swk\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868712 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-env-overrides\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868729 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-socket-dir-parent\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868795 5110 reconciler_common.go:299] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868807 5110 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868818 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868829 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868840 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868850 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868860 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868871 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l87hs\" (UniqueName: \"kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868882 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-w94wk\" (UniqueName: \"kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868892 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868902 5110 reconciler_common.go:299] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868913 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868924 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868933 5110 reconciler_common.go:299] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868943 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868954 5110 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868965 5110 reconciler_common.go:299] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868976 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868986 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d4tqq\" (UniqueName: \"kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.868997 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nb9c\" (UniqueName: \"kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869007 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869017 5110 reconciler_common.go:299] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869028 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869039 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869049 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869086 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869105 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xnxbn\" (UniqueName: \"kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869120 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869130 5110 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869142 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tkdh6\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869152 5110 reconciler_common.go:299] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869164 5110 reconciler_common.go:299] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869174 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869185 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869271 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m5g58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baee355c-9055-42f1-87c4-3dc7a08b68f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnc69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnc69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m5g58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.869739 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.870178 5110 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.870239 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:12.370222509 +0000 UTC m=+96.402834021 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.871704 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca" (OuterVolumeSpecName: "service-ca") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.872170 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp" (OuterVolumeSpecName: "tmp") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.872219 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities" (OuterVolumeSpecName: "utilities") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.872794 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert" (OuterVolumeSpecName: "cert") pod "a52afe44-fb37-46ed-a1f8-bf39727a3cbe" (UID: "a52afe44-fb37-46ed-a1f8-bf39727a3cbe"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.873241 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "42a11a02-47e1-488f-b270-2679d3298b0e" (UID: "42a11a02-47e1-488f-b270-2679d3298b0e"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.873254 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.873638 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.873742 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw" (OuterVolumeSpecName: "kube-api-access-9z4sw") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "kube-api-access-9z4sw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.873919 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874010 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem" (OuterVolumeSpecName: "ca-trust-extracted-pem") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "ca-trust-extracted-pem". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874126 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7" (OuterVolumeSpecName: "kube-api-access-hm9x7") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "kube-api-access-hm9x7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874285 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874150 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config" (OuterVolumeSpecName: "config") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874593 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "f7e2c886-118e-43bb-bef1-c78134de392b" (UID: "f7e2c886-118e-43bb-bef1-c78134de392b"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874876 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874908 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.874917 5110 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.875266 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.875399 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config" (OuterVolumeSpecName: "config") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.875680 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key" (OuterVolumeSpecName: "signing-key") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.876023 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf" (OuterVolumeSpecName: "kube-api-access-q4smf") pod "0dd0fbac-8c0d-4228-8faa-abbeedabf7db" (UID: "0dd0fbac-8c0d-4228-8faa-abbeedabf7db"). InnerVolumeSpecName "kube-api-access-q4smf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.876171 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv" (OuterVolumeSpecName: "kube-api-access-dztfv") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "kube-api-access-dztfv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.876525 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz" (OuterVolumeSpecName: "kube-api-access-ws8zz") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "kube-api-access-ws8zz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.876871 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.877640 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf" (OuterVolumeSpecName: "kube-api-access-6dmhf") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "kube-api-access-6dmhf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.878392 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr" (OuterVolumeSpecName: "kube-api-access-6g4lr") pod "f7e2c886-118e-43bb-bef1-c78134de392b" (UID: "f7e2c886-118e-43bb-bef1-c78134de392b"). InnerVolumeSpecName "kube-api-access-6g4lr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.879393 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps" (OuterVolumeSpecName: "kube-api-access-d7cps") pod "af41de71-79cf-4590-bbe9-9e8b848862cb" (UID: "af41de71-79cf-4590-bbe9-9e8b848862cb"). InnerVolumeSpecName "kube-api-access-d7cps". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.879492 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.879847 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.879993 5110 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.880482 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:12.380456884 +0000 UTC m=+96.413068426 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.880189 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.881110 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr" (OuterVolumeSpecName: "kube-api-access-wj4qr") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "kube-api-access-wj4qr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.881140 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.881331 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk" (OuterVolumeSpecName: "kube-api-access-qqbfk") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "kube-api-access-qqbfk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.881896 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6" (OuterVolumeSpecName: "kube-api-access-pllx6") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "kube-api-access-pllx6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.882115 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.882310 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "a208c9c2-333b-4b4a-be0d-bc32ec38a821" (UID: "a208c9c2-333b-4b4a-be0d-bc32ec38a821"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.882384 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.882438 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.882940 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc" (OuterVolumeSpecName: "kube-api-access-zg8nc") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "kube-api-access-zg8nc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.883128 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv" (OuterVolumeSpecName: "kube-api-access-6rmnv") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "kube-api-access-6rmnv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.883295 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m" (OuterVolumeSpecName: "kube-api-access-4hb7m") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "kube-api-access-4hb7m". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.883351 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca" (OuterVolumeSpecName: "service-ca") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.883559 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.884312 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.885160 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-etcd/etcd-crc" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.883787 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.883849 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.884913 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.886278 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.886394 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.886470 5110 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.886486 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw" (OuterVolumeSpecName: "kube-api-access-5lcfw") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "kube-api-access-5lcfw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.885473 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.885825 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.885328 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9" (OuterVolumeSpecName: "kube-api-access-ddlk9") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "kube-api-access-ddlk9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.886668 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities" (OuterVolumeSpecName: "utilities") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.886879 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:12.386579614 +0000 UTC m=+96.419191136 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.887083 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl" (OuterVolumeSpecName: "kube-api-access-twvbl") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "kube-api-access-twvbl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.887622 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.887970 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config" (OuterVolumeSpecName: "config") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.888605 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.888851 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.888916 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd" (OuterVolumeSpecName: "kube-api-access-8pskd") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "kube-api-access-8pskd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.889176 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:12.389162446 +0000 UTC m=+96.421773968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.889375 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config" (OuterVolumeSpecName: "config") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.889667 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.890501 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2" (OuterVolumeSpecName: "kube-api-access-ks6v2") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "kube-api-access-ks6v2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.890992 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.890990 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l" (OuterVolumeSpecName: "kube-api-access-sbc2l") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "kube-api-access-sbc2l". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.891257 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.891399 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.891748 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.891986 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.892232 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.892375 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.892703 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit" (OuterVolumeSpecName: "audit") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.892729 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.893139 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.893259 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894006 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894033 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.893359 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h" (OuterVolumeSpecName: "kube-api-access-94l9h") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "kube-api-access-94l9h". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894034 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894135 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s" (OuterVolumeSpecName: "kube-api-access-xfp5s") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "kube-api-access-xfp5s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.893619 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.893835 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9" (OuterVolumeSpecName: "kube-api-access-99zj9") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "kube-api-access-99zj9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894138 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894599 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv" (OuterVolumeSpecName: "kube-api-access-pddnv") pod "e093be35-bb62-4843-b2e8-094545761610" (UID: "e093be35-bb62-4843-b2e8-094545761610"). InnerVolumeSpecName "kube-api-access-pddnv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894728 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" (UID: "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894878 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp" (OuterVolumeSpecName: "tmp") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.894947 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg" (OuterVolumeSpecName: "kube-api-access-wbmqg") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "kube-api-access-wbmqg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895001 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895016 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895448 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895685 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq" (OuterVolumeSpecName: "kube-api-access-m26jq") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "kube-api-access-m26jq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895902 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895752 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config" (OuterVolumeSpecName: "config") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.896624 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.896651 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap" (OuterVolumeSpecName: "whereabouts-flatfile-configmap") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "whereabouts-flatfile-configmap". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.897103 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.897127 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.897140 5110 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.897168 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca" (OuterVolumeSpecName: "client-ca") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.897197 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:12.397180889 +0000 UTC m=+96.429792411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.896129 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca" (OuterVolumeSpecName: "serviceca") pod "5ebfebf6-3ecd-458e-943f-bb25b52e2718" (UID: "5ebfebf6-3ecd-458e-943f-bb25b52e2718"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.896178 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t" (OuterVolumeSpecName: "kube-api-access-zth6t") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "kube-api-access-zth6t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.895697 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config" (OuterVolumeSpecName: "config") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.897334 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.899261 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume" (OuterVolumeSpecName: "config-volume") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.900185 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.900275 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.900322 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.900502 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "0dd0fbac-8c0d-4228-8faa-abbeedabf7db" (UID: "0dd0fbac-8c0d-4228-8faa-abbeedabf7db"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.900815 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60643394-bb94-4b51-a9c7-883e2359977a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e19e709ef87fb9cb61061fe5d588545d4e6145cb814255fd180b74cae14a991f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f91a3deb02a02a3162c98c11c3614f0692f7ccc8366e5de75ac9d079142ac507\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://53b7011b6f8c381cac7dfbb307f74d4a1bc3c33189e417c20d240acaf720cec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.900943 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.901535 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.902051 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs" (OuterVolumeSpecName: "certs") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.903395 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config" (OuterVolumeSpecName: "config") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.903573 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities" (OuterVolumeSpecName: "utilities") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.905422 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images" (OuterVolumeSpecName: "images") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.905908 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.905997 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.906079 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.906149 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.906208 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:11Z","lastTransitionTime":"2026-03-17T18:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.905948 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.907620 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.908223 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj" (OuterVolumeSpecName: "kube-api-access-qgrkj") pod "42a11a02-47e1-488f-b270-2679d3298b0e" (UID: "42a11a02-47e1-488f-b270-2679d3298b0e"). InnerVolumeSpecName "kube-api-access-qgrkj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.908883 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz" (OuterVolumeSpecName: "kube-api-access-grwfz") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "kube-api-access-grwfz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.909168 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr" (OuterVolumeSpecName: "kube-api-access-z5rsr") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "kube-api-access-z5rsr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.909311 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.909418 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.909607 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.909995 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.910490 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9" (OuterVolumeSpecName: "kube-api-access-9vsz9") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "kube-api-access-9vsz9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.910820 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.910900 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd" (OuterVolumeSpecName: "kube-api-access-mjwtd") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "kube-api-access-mjwtd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.910916 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.911292 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.911371 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images" (OuterVolumeSpecName: "images") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.912754 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nt2j\" (UniqueName: \"kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.912888 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.913367 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config" (OuterVolumeSpecName: "config") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.913430 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.913445 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.913595 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.913655 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.913631 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.912736 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh" (OuterVolumeSpecName: "kube-api-access-m5lgh") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "kube-api-access-m5lgh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.914414 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv" (OuterVolumeSpecName: "kube-api-access-xxfcv") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "kube-api-access-xxfcv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.915366 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.915932 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.916713 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7xz2\" (UniqueName: \"kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.916725 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgwk\" (UniqueName: \"kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.920838 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.923500 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.925410 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.926274 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.934120 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.942985 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-slqfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-slqfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.943437 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.951676 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.954656 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.957356 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4jphj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.966377 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2nhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2nhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-jr488\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970272 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-ovn\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970395 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-netns\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970493 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6gfq8\" (UniqueName: \"kubernetes.io/projected/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-kube-api-access-6gfq8\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970570 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-systemd-units\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970646 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-slash\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970729 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970807 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-multus-certs\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970911 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-host\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.970981 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-netd\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971067 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-cnibin\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971137 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-cni-multus\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971204 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971282 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fnc69\" (UniqueName: \"kubernetes.io/projected/baee355c-9055-42f1-87c4-3dc7a08b68f6-kube-api-access-fnc69\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971350 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7da49224-3b49-41d3-8490-ae2724128e67-rootfs\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971420 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-config\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971487 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-etc-kubernetes\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971554 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-system-cni-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971625 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7da49224-3b49-41d3-8490-ae2724128e67-mcd-auth-proxy-config\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971692 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-serviceca\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971878 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-netd\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.971926 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7da49224-3b49-41d3-8490-ae2724128e67-rootfs\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972007 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-ovn\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972019 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-cnibin\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972047 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-cni-multus\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972140 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972210 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-multus-certs\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972287 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-host\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972643 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-system-cni-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972665 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-slash\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972701 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-etc-kubernetes\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972876 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d2nhm\" (UniqueName: \"kubernetes.io/projected/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-kube-api-access-d2nhm\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.972953 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-var-lib-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973031 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-node-log\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973124 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-var-lib-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973142 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-node-log\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973134 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovn-node-metrics-cert\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973125 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-systemd-units\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973006 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-netns\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973314 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7346b312-05b5-4475-8d54-82709b69e7ed-cni-binary-copy\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973345 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-k8s-cni-cncf-io\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973512 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-run-k8s-cni-cncf-io\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973589 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980117 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980238 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-conf-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980336 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980424 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980506 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980483 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68ae3b0b-110f-4649-af0a-84e9a2a49817\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://892b0c9c14fabad6da32657ff71c3a8b34d373f96898fd63f4c538fc11e3552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:38Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://deff5a49b7611cfa357b185827a99ebb293931b96fcd7303036b7a26b5b350c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:38Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://8686316b4620286368e9ae51f21b41341ab76b0a6173743fbb9345e41cb8cfc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980596 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-log-socket\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980877 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-hosts-file\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980921 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-system-cni-dir\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980957 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-os-release\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.980988 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-ovn-kubernetes\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981023 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-script-lib\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981084 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981117 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-systemd\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981164 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-cni-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981195 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-os-release\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981225 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7346b312-05b5-4475-8d54-82709b69e7ed-multus-daemon-config\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981256 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-bin\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981332 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-tmp-dir\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981364 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7da49224-3b49-41d3-8490-ae2724128e67-proxy-tls\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981395 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981433 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lck5w\" (UniqueName: \"kubernetes.io/projected/7346b312-05b5-4475-8d54-82709b69e7ed-kube-api-access-lck5w\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981466 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981498 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-csf4c\" (UniqueName: \"kubernetes.io/projected/7da49224-3b49-41d3-8490-ae2724128e67-kube-api-access-csf4c\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981531 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-cni-bin\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981561 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-kubelet\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981592 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-hostroot\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981624 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-stxn5\" (UniqueName: \"kubernetes.io/projected/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-kube-api-access-stxn5\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981654 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cnibin\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981682 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cni-binary-copy\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981713 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8b652\" (UniqueName: \"kubernetes.io/projected/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-kube-api-access-8b652\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981747 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981781 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85swk\" (UniqueName: \"kubernetes.io/projected/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-kube-api-access-85swk\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981813 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-env-overrides\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981844 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-socket-dir-parent\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981876 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-kubelet\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981907 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-netns\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.981939 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-etc-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982096 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982117 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l9stx\" (UniqueName: \"kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982136 5110 reconciler_common.go:299] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982157 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nspp\" (UniqueName: \"kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982176 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982193 5110 reconciler_common.go:299] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982209 5110 reconciler_common.go:299] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982226 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982243 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982260 5110 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982277 5110 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982293 5110 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982311 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hm9x7\" (UniqueName: \"kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982327 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982345 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dztfv\" (UniqueName: \"kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982346 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-serviceca\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982362 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982378 5110 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982396 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nmmzf\" (UniqueName: \"kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982414 5110 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982426 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-systemd\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982442 5110 reconciler_common.go:299] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982460 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rzt4w\" (UniqueName: \"kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982472 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-hosts-file\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982478 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982494 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982501 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-system-cni-dir\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982511 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4g8ts\" (UniqueName: \"kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982528 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ptkcf\" (UniqueName: \"kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982546 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982555 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-os-release\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982563 5110 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982578 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982584 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-ovn-kubernetes\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.982595 5110 reconciler_common.go:299] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.974161 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973763 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.973846 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7da49224-3b49-41d3-8490-ae2724128e67-mcd-auth-proxy-config\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.974309 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-config\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.974458 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7346b312-05b5-4475-8d54-82709b69e7ed-cni-binary-copy\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.983937 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-log-socket\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.986545 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.992305 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-netns\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.992326 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-kubelet\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.992553 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-host-var-lib-cni-bin\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.992630 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-conf-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.992930 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-env-overrides\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993036 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993299 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7346b312-05b5-4475-8d54-82709b69e7ed-multus-daemon-config\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993347 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-hostroot\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993374 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993408 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-etc-openvswitch\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993456 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-tmp-dir\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.993469 5110 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: E0317 18:49:11.993536 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs podName:baee355c-9055-42f1-87c4-3dc7a08b68f6 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:12.493518859 +0000 UTC m=+96.526130381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs") pod "network-metrics-daemon-m5g58" (UID: "baee355c-9055-42f1-87c4-3dc7a08b68f6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993694 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-os-release\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993727 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cnibin\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993812 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993847 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.993956 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-bin\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.994417 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.994977 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cni-binary-copy\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.995026 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-script-lib\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.995638 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-socket-dir-parent\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.995787 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7346b312-05b5-4475-8d54-82709b69e7ed-multus-cni-dir\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.995815 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-kubelet\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.995966 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996037 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996064 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996074 5110 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996082 5110 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996090 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996100 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zsb9b\" (UniqueName: \"kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996110 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7jjkz\" (UniqueName: \"kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996120 5110 reconciler_common.go:299] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996128 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996137 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ftwb6\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996145 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996153 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tknt7\" (UniqueName: \"kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996151 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gfq8\" (UniqueName: \"kubernetes.io/projected/ab807a2a-8744-4ebe-8e74-f841ffbdac5e-kube-api-access-6gfq8\") pod \"node-ca-slqfs\" (UID: \"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\") " pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996162 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996333 5110 reconciler_common.go:299] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996352 5110 reconciler_common.go:299] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996367 5110 reconciler_common.go:299] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996380 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996392 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996405 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996417 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996430 5110 reconciler_common.go:299] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996442 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996455 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996468 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hckvg\" (UniqueName: \"kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996490 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-26xrl\" (UniqueName: \"kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996504 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qqbfk\" (UniqueName: \"kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996517 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996530 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996543 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996555 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996567 5110 reconciler_common.go:299] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996579 5110 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996591 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996605 5110 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996617 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996629 5110 reconciler_common.go:299] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996642 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9z4sw\" (UniqueName: \"kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996657 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996670 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mfzkj\" (UniqueName: \"kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996683 5110 reconciler_common.go:299] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996695 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996705 5110 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996718 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996730 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q4smf\" (UniqueName: \"kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996743 5110 reconciler_common.go:299] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996756 5110 reconciler_common.go:299] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996802 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2nhm\" (UniqueName: \"kubernetes.io/projected/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-kube-api-access-d2nhm\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996805 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7da49224-3b49-41d3-8490-ae2724128e67-proxy-tls\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.997296 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovn-node-metrics-cert\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.996769 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pllx6\" (UniqueName: \"kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998322 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4hb7m\" (UniqueName: \"kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998336 5110 reconciler_common.go:299] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998350 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998394 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-twvbl\" (UniqueName: \"kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998408 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998421 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998433 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998468 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998482 5110 reconciler_common.go:299] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998495 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998507 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998521 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998555 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ks6v2\" (UniqueName: \"kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998570 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998583 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998596 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998631 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998660 5110 reconciler_common.go:299] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998673 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998687 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8pskd\" (UniqueName: \"kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998724 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998736 5110 reconciler_common.go:299] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998749 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998761 5110 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998795 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998808 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998821 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998833 5110 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998846 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m5lgh\" (UniqueName: \"kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998882 5110 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998896 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mjwtd\" (UniqueName: \"kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998909 5110 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998922 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998955 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xxfcv\" (UniqueName: \"kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998970 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998985 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.998999 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ws8zz\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999012 5110 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999045 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d7cps\" (UniqueName: \"kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999088 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6g4lr\" (UniqueName: \"kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999102 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wj4qr\" (UniqueName: \"kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999115 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999128 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zg8nc\" (UniqueName: \"kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999162 5110 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999176 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ddlk9\" (UniqueName: \"kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999189 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sbc2l\" (UniqueName: \"kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999201 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999213 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999249 5110 reconciler_common.go:299] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999262 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-z5rsr\" (UniqueName: \"kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999276 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qgrkj\" (UniqueName: \"kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999288 5110 reconciler_common.go:299] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999322 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999337 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zth6t\" (UniqueName: \"kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999349 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999440 5110 reconciler_common.go:299] "Volume detached for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999456 5110 reconciler_common.go:299] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999492 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9vsz9\" (UniqueName: \"kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999506 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999519 5110 reconciler_common.go:299] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:11 crc kubenswrapper[5110]: I0317 18:49:11.999531 5110 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999500 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999544 5110 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999579 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-94l9h\" (UniqueName: \"kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999592 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999604 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xfp5s\" (UniqueName: \"kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999619 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wbmqg\" (UniqueName: \"kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999657 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999671 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pddnv\" (UniqueName: \"kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999684 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999697 5110 reconciler_common.go:299] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999732 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999748 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m26jq\" (UniqueName: \"kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999761 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999773 5110 reconciler_common.go:299] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999787 5110 reconciler_common.go:299] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999823 5110 reconciler_common.go:299] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999838 5110 reconciler_common.go:299] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999851 5110 reconciler_common.go:299] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999864 5110 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999901 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999916 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999930 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6rmnv\" (UniqueName: \"kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999943 5110 reconciler_common.go:299] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999977 5110 reconciler_common.go:299] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:11.999992 5110 reconciler_common.go:299] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000006 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000019 5110 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000031 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000076 5110 reconciler_common.go:299] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000089 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000103 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000138 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000151 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000164 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6dmhf\" (UniqueName: \"kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000177 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-99zj9\" (UniqueName: \"kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000189 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000225 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5lcfw\" (UniqueName: \"kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000241 5110 reconciler_common.go:299] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000255 5110 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000268 5110 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000302 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-grwfz\" (UniqueName: \"kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000315 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000328 5110 reconciler_common.go:299] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000341 5110 reconciler_common.go:299] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000352 5110 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000387 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pgx6b\" (UniqueName: \"kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000400 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000412 5110 reconciler_common.go:299] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000424 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000458 5110 reconciler_common.go:299] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000472 5110 reconciler_common.go:299] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.000484 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.004470 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-jr488\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.005118 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnc69\" (UniqueName: \"kubernetes.io/projected/baee355c-9055-42f1-87c4-3dc7a08b68f6-kube-api-access-fnc69\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.009550 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b652\" (UniqueName: \"kubernetes.io/projected/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-kube-api-access-8b652\") pod \"ovnkube-node-4jphj\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.009762 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.009790 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.009801 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.009817 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.009836 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.010046 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lck5w\" (UniqueName: \"kubernetes.io/projected/7346b312-05b5-4475-8d54-82709b69e7ed-kube-api-access-lck5w\") pod \"multus-6kh2q\" (UID: \"7346b312-05b5-4475-8d54-82709b69e7ed\") " pod="openshift-multus/multus-6kh2q" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.011454 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-stxn5\" (UniqueName: \"kubernetes.io/projected/d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3-kube-api-access-stxn5\") pod \"node-resolver-6k4l7\" (UID: \"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\") " pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.016034 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-85swk\" (UniqueName: \"kubernetes.io/projected/341f2a23-d987-42e8-8ca7-f4f596ef3f3d-kube-api-access-85swk\") pod \"multus-additional-cni-plugins-lh6lx\" (UID: \"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\") " pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.016122 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-csf4c\" (UniqueName: \"kubernetes.io/projected/7da49224-3b49-41d3-8490-ae2724128e67-kube-api-access-csf4c\") pod \"machine-config-daemon-6jbgs\" (UID: \"7da49224-3b49-41d3-8490-ae2724128e67\") " pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.016474 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.023790 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-6k4l7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stxn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6k4l7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.025327 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.033207 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-6kh2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7346b312-05b5-4475-8d54-82709b69e7ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lck5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6kh2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.041017 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34177974_8d82_49d2_a763_391d0df3bbd8.slice/crio-29f89eab423d8279433aba95f15c10fb7f0a42f06d069c08aca9bfb09dfee09a WatchSource:0}: Error finding container 29f89eab423d8279433aba95f15c10fb7f0a42f06d069c08aca9bfb09dfee09a: Status 404 returned error can't find the container with id 29f89eab423d8279433aba95f15c10fb7f0a42f06d069c08aca9bfb09dfee09a Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.053418 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.065135 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.075528 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod428b39f5_eb1c_4f65_b7a4_eeb6e84860cc.slice/crio-926716b76042dadf4684a4a4655a41b34b0f937f3aa92b60f8f5fb566bb29fbf WatchSource:0}: Error finding container 926716b76042dadf4684a4a4655a41b34b0f937f3aa92b60f8f5fb566bb29fbf: Status 404 returned error can't find the container with id 926716b76042dadf4684a4a4655a41b34b0f937f3aa92b60f8f5fb566bb29fbf Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.081296 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6k4l7" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.092569 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-slqfs" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.103094 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6kh2q" Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.103205 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0fce2ad_638a_4ad5_afc3_3cd9c4f125e3.slice/crio-cfe808781a324b01a3958e2f88d4598b8e3f4bc37e628ef76dc14db5a9fc1718 WatchSource:0}: Error finding container cfe808781a324b01a3958e2f88d4598b8e3f4bc37e628ef76dc14db5a9fc1718: Status 404 returned error can't find the container with id cfe808781a324b01a3958e2f88d4598b8e3f4bc37e628ef76dc14db5a9fc1718 Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.110483 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.112006 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.112069 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.112085 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.112104 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.112116 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.118552 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.125015 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.127010 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7346b312_05b5_4475_8d54_82709b69e7ed.slice/crio-7eb889795e9c9dc8f77804121bdcc0bdddcadd491163a5aae2778457f67c0819 WatchSource:0}: Error finding container 7eb889795e9c9dc8f77804121bdcc0bdddcadd491163a5aae2778457f67c0819: Status 404 returned error can't find the container with id 7eb889795e9c9dc8f77804121bdcc0bdddcadd491163a5aae2778457f67c0819 Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.128206 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab807a2a_8744_4ebe_8e74_f841ffbdac5e.slice/crio-a657567637b89c79814e199a726031573c55533ebc8fd2ee74baf840aa9f6f94 WatchSource:0}: Error finding container a657567637b89c79814e199a726031573c55533ebc8fd2ee74baf840aa9f6f94: Status 404 returned error can't find the container with id a657567637b89c79814e199a726031573c55533ebc8fd2ee74baf840aa9f6f94 Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.131732 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.164641 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod341f2a23_d987_42e8_8ca7_f4f596ef3f3d.slice/crio-abe8d290bea4a935fa271e4deb874200be055b2425c4a0f6c1385c905130b88b WatchSource:0}: Error finding container abe8d290bea4a935fa271e4deb874200be055b2425c4a0f6c1385c905130b88b: Status 404 returned error can't find the container with id abe8d290bea4a935fa271e4deb874200be055b2425c4a0f6c1385c905130b88b Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.168771 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf20fbc8f_f15b_4713_9b57_aa5dbcb0d9b8.slice/crio-d9fdf49e963657295a75c522c5fb9bc94198871a002c11e3280ef9eb82012cdc WatchSource:0}: Error finding container d9fdf49e963657295a75c522c5fb9bc94198871a002c11e3280ef9eb82012cdc: Status 404 returned error can't find the container with id d9fdf49e963657295a75c522c5fb9bc94198871a002c11e3280ef9eb82012cdc Mar 17 18:49:12 crc kubenswrapper[5110]: W0317 18:49:12.190736 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe3aacec_3e82_49dc_833e_4b1d848ea2c6.slice/crio-98f18290f31e5aaa7954e356a85a85b8e221a7703ec1956532614cb7401153aa WatchSource:0}: Error finding container 98f18290f31e5aaa7954e356a85a85b8e221a7703ec1956532614cb7401153aa: Status 404 returned error can't find the container with id 98f18290f31e5aaa7954e356a85a85b8e221a7703ec1956532614cb7401153aa Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.230725 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.230960 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.231009 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.231026 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.231035 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.335611 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.335641 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.335649 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.335663 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.335672 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.405217 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.405308 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.405331 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.405373 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.405389 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405436 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:13.405399836 +0000 UTC m=+97.438011358 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405509 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405524 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405534 5110 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405535 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405555 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405567 5110 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405581 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:13.40556684 +0000 UTC m=+97.438178362 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405611 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:13.405599821 +0000 UTC m=+97.438211343 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405630 5110 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405661 5110 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405682 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:13.405670033 +0000 UTC m=+97.438281565 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.405771 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:13.405748715 +0000 UTC m=+97.438360237 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.445859 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.445913 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.445926 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.445941 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.445987 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.478497 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" event={"ID":"fe3aacec-3e82-49dc-833e-4b1d848ea2c6","Type":"ContainerStarted","Data":"98f18290f31e5aaa7954e356a85a85b8e221a7703ec1956532614cb7401153aa"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.479610 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6kh2q" event={"ID":"7346b312-05b5-4475-8d54-82709b69e7ed","Type":"ContainerStarted","Data":"707281b7717436044e263464eec7435119e33d38877132718dd38f6a8a0acdbb"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.479632 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6kh2q" event={"ID":"7346b312-05b5-4475-8d54-82709b69e7ed","Type":"ContainerStarted","Data":"7eb889795e9c9dc8f77804121bdcc0bdddcadd491163a5aae2778457f67c0819"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.481336 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"85a196228f75ca3c72b35db9097cedcadf214897cd670aaa987893308b51fadf"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.481358 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"aff0bd0446f9a414a233b7615630637489d211049e6faa7e85e89ea14366b7f9"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.481367 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"5a2d08ccb4cea6bea78fdb7b591edb6e43f84f45226962ff5ce2ce05d1e18c9f"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.482424 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" event={"ID":"34177974-8d82-49d2-a763-391d0df3bbd8","Type":"ContainerStarted","Data":"ce0d9124d558d61880506fe9717799b575efe149b547859b524cd915c8210c9c"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.482446 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" event={"ID":"34177974-8d82-49d2-a763-391d0df3bbd8","Type":"ContainerStarted","Data":"29f89eab423d8279433aba95f15c10fb7f0a42f06d069c08aca9bfb09dfee09a"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.483575 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"0075007a0a6d1dcc16449135cf348101cd8615648447feb043c5978c7d129075"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.483594 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"c16c365ecf8a42057b8434c5452aa7ec7cb690c20b64cac610c4332a31226478"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.484697 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" exitCode=0 Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.484746 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.484759 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"d9fdf49e963657295a75c522c5fb9bc94198871a002c11e3280ef9eb82012cdc"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.503022 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerStarted","Data":"abe8d290bea4a935fa271e4deb874200be055b2425c4a0f6c1385c905130b88b"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.504293 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6k4l7" event={"ID":"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3","Type":"ContainerStarted","Data":"cfe808781a324b01a3958e2f88d4598b8e3f4bc37e628ef76dc14db5a9fc1718"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.506449 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.506684 5110 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.506772 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs podName:baee355c-9055-42f1-87c4-3dc7a08b68f6 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:13.506749565 +0000 UTC m=+97.539361087 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs") pod "network-metrics-daemon-m5g58" (UID: "baee355c-9055-42f1-87c4-3dc7a08b68f6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.509765 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-slqfs" event={"ID":"ab807a2a-8744-4ebe-8e74-f841ffbdac5e","Type":"ContainerStarted","Data":"a657567637b89c79814e199a726031573c55533ebc8fd2ee74baf840aa9f6f94"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.513941 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60643394-bb94-4b51-a9c7-883e2359977a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e19e709ef87fb9cb61061fe5d588545d4e6145cb814255fd180b74cae14a991f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f91a3deb02a02a3162c98c11c3614f0692f7ccc8366e5de75ac9d079142ac507\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://53b7011b6f8c381cac7dfbb307f74d4a1bc3c33189e417c20d240acaf720cec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.514266 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" event={"ID":"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc","Type":"ContainerStarted","Data":"926716b76042dadf4684a4a4655a41b34b0f937f3aa92b60f8f5fb566bb29fbf"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.524660 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.536987 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.548613 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.549541 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.549581 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.549597 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.549616 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.549628 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.562212 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-slqfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-slqfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.578065 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4jphj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.590122 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2nhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2nhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-jr488\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.608273 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68ae3b0b-110f-4649-af0a-84e9a2a49817\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://892b0c9c14fabad6da32657ff71c3a8b34d373f96898fd63f4c538fc11e3552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:38Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://deff5a49b7611cfa357b185827a99ebb293931b96fcd7303036b7a26b5b350c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:38Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://8686316b4620286368e9ae51f21b41341ab76b0a6173743fbb9345e41cb8cfc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.619611 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.629280 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.639349 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-6k4l7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stxn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6k4l7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.651742 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-6kh2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7346b312-05b5-4475-8d54-82709b69e7ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"65Mi\\\"},\\\"containerID\\\":\\\"cri-o://707281b7717436044e263464eec7435119e33d38877132718dd38f6a8a0acdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"65Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:49:12Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lck5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6kh2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.663769 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.663805 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.663814 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.663830 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.663839 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.663775 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lh6lx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.682370 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7da49224-3b49-41d3-8490-ae2724128e67\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-csf4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-csf4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6jbgs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.708456 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b555f50b-3fcc-4241-b747-d8c898b70530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://6f8f146b9f6264d1b969799f87c68b356d3d701c5836b314a29278bd223e63ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://18527c5fa347f446246a519571a862dc05435325ff8bb4de538c7834fa60f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://638c38a4a2517c3d40ecaeafeb3aff8c788d7679eec37d03bbf83b1d3227cdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e30ed6978ad51bf9888576bae1beb1a9c545283b784e4b0aa1d996642a002616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:42Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://bfcb8305c913d0503cc502dcbfbdedb73b129c8d9b557cf913169eda6efd6d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://a29edae24756ec647faeaacc3417fe2dc3d9edbd4dd08de37d3d930ea9060a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a29edae24756ec647faeaacc3417fe2dc3d9edbd4dd08de37d3d930ea9060a00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://652b9629cee1d27b3cd8bb6186c5d479f4614b794bd8b255c10e38cc41d4ac0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652b9629cee1d27b3cd8bb6186c5d479f4614b794bd8b255c10e38cc41d4ac0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://65cf9a08bb479461bfc81b04b4184f9881e3354555c9a3b0c77409d162e9b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65cf9a08bb479461bfc81b04b4184f9881e3354555c9a3b0c77409d162e9b15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.722404 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.733423 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7ff9391-37cf-48cb-91a1-2ce430fbc515\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://8545c60fc2c493fe31c8c9c36d9eac4f122e0feeea1c08863a4e28302fa86510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://13c01ebb653a6bdaad86850e3714878a54746b6c7cc793fd32d45dade13e7e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13c01ebb653a6bdaad86850e3714878a54746b6c7cc793fd32d45dade13e7e85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.741914 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331f5282-47de-4f21-903e-5dfb189879f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T18:48:49Z\\\",\\\"message\\\":\\\" envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"ClientsAllowCBOR\\\\\\\" enabled=false\\\\nW0317 18:48:49.126497 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 18:48:49.126663 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI0317 18:48:49.127547 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-97051896/tls.crt::/tmp/serving-cert-97051896/tls.key\\\\\\\"\\\\nI0317 18:48:49.391789 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 18:48:49.393987 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 18:48:49.394005 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 18:48:49.394028 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 18:48:49.394034 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 18:48:49.397362 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 18:48:49.397406 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0317 18:48:49.397405 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0317 18:48:49.397411 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 18:48:49.397432 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 18:48:49.397438 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 18:48:49.397442 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 18:48:49.397445 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0317 18:48:49.399325 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T18:48:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.743891 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:12 crc kubenswrapper[5110]: E0317 18:49:12.743968 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.746695 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m5g58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baee355c-9055-42f1-87c4-3dc7a08b68f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnc69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnc69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m5g58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.753193 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2nhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d2nhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-jr488\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.760201 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68ae3b0b-110f-4649-af0a-84e9a2a49817\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://892b0c9c14fabad6da32657ff71c3a8b34d373f96898fd63f4c538fc11e3552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:38Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://deff5a49b7611cfa357b185827a99ebb293931b96fcd7303036b7a26b5b350c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:38Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://8686316b4620286368e9ae51f21b41341ab76b0a6173743fbb9345e41cb8cfc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.765944 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.770791 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.770817 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.770826 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.770838 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.770847 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.771528 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01080b46-74f1-4191-8755-5152a57b3b25" path="/var/lib/kubelet/pods/01080b46-74f1-4191-8755-5152a57b3b25/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.772774 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09cfa50b-4138-4585-a53e-64dd3ab73335" path="/var/lib/kubelet/pods/09cfa50b-4138-4585-a53e-64dd3ab73335/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.774797 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" path="/var/lib/kubelet/pods/0dd0fbac-8c0d-4228-8faa-abbeedabf7db/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.776894 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0effdbcf-dd7d-404d-9d48-77536d665a5d" path="/var/lib/kubelet/pods/0effdbcf-dd7d-404d-9d48-77536d665a5d/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.778800 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://85a196228f75ca3c72b35db9097cedcadf214897cd670aaa987893308b51fadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:49:12Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0,1000500000],\\\"uid\\\":1000500000}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://aff0bd0446f9a414a233b7615630637489d211049e6faa7e85e89ea14366b7f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:49:12Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0,1000500000],\\\"uid\\\":1000500000}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.779509 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="149b3c48-e17c-4a66-a835-d86dabf6ff13" path="/var/lib/kubelet/pods/149b3c48-e17c-4a66-a835-d86dabf6ff13/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.785576 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16bdd140-dce1-464c-ab47-dd5798d1d256" path="/var/lib/kubelet/pods/16bdd140-dce1-464c-ab47-dd5798d1d256/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.786355 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f80adb-c1c3-49ba-8ee4-932c851d3897" path="/var/lib/kubelet/pods/18f80adb-c1c3-49ba-8ee4-932c851d3897/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.786376 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-6k4l7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stxn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6k4l7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.787781 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" path="/var/lib/kubelet/pods/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.788466 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2325ffef-9d5b-447f-b00e-3efc429acefe" path="/var/lib/kubelet/pods/2325ffef-9d5b-447f-b00e-3efc429acefe/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.789831 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="301e1965-1754-483d-b6cc-bfae7038bbca" path="/var/lib/kubelet/pods/301e1965-1754-483d-b6cc-bfae7038bbca/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.791984 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31fa8943-81cc-4750-a0b7-0fa9ab5af883" path="/var/lib/kubelet/pods/31fa8943-81cc-4750-a0b7-0fa9ab5af883/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.793535 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a11a02-47e1-488f-b270-2679d3298b0e" path="/var/lib/kubelet/pods/42a11a02-47e1-488f-b270-2679d3298b0e/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.794160 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567683bd-0efc-4f21-b076-e28559628404" path="/var/lib/kubelet/pods/567683bd-0efc-4f21-b076-e28559628404/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.796174 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584e1f4a-8205-47d7-8efb-3afc6017c4c9" path="/var/lib/kubelet/pods/584e1f4a-8205-47d7-8efb-3afc6017c4c9/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.796405 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-6kh2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7346b312-05b5-4475-8d54-82709b69e7ed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"65Mi\\\"},\\\"containerID\\\":\\\"cri-o://707281b7717436044e263464eec7435119e33d38877132718dd38f6a8a0acdbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"65Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:49:12Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lck5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6kh2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.796613 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593a3561-7760-45c5-8f91-5aaef7475d0f" path="/var/lib/kubelet/pods/593a3561-7760-45c5-8f91-5aaef7475d0f/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.797860 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ebfebf6-3ecd-458e-943f-bb25b52e2718" path="/var/lib/kubelet/pods/5ebfebf6-3ecd-458e-943f-bb25b52e2718/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.798585 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6077b63e-53a2-4f96-9d56-1ce0324e4913" path="/var/lib/kubelet/pods/6077b63e-53a2-4f96-9d56-1ce0324e4913/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.799770 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" path="/var/lib/kubelet/pods/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.801172 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6edfcf45-925b-4eff-b940-95b6fc0b85d4" path="/var/lib/kubelet/pods/6edfcf45-925b-4eff-b940-95b6fc0b85d4/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.802601 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ee8fbd3-1f81-4666-96da-5afc70819f1a" path="/var/lib/kubelet/pods/6ee8fbd3-1f81-4666-96da-5afc70819f1a/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.806883 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" path="/var/lib/kubelet/pods/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.865772 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736c54fe-349c-4bb9-870a-d1c1d1c03831" path="/var/lib/kubelet/pods/736c54fe-349c-4bb9-870a-d1c1d1c03831/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.867551 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7599e0b6-bddf-4def-b7f2-0b32206e8651" path="/var/lib/kubelet/pods/7599e0b6-bddf-4def-b7f2-0b32206e8651/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.869659 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7afa918d-be67-40a6-803c-d3b0ae99d815" path="/var/lib/kubelet/pods/7afa918d-be67-40a6-803c-d3b0ae99d815/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.870476 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df94c10-441d-4386-93a6-6730fb7bcde0" path="/var/lib/kubelet/pods/7df94c10-441d-4386-93a6-6730fb7bcde0/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.871774 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" path="/var/lib/kubelet/pods/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.874524 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341f2a23-d987-42e8-8ca7-f4f596ef3f3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85swk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lh6lx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.875729 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.875763 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.875773 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.875786 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.875802 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.881797 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e39f7b-62e4-4fc9-992a-6535ce127a02" path="/var/lib/kubelet/pods/81e39f7b-62e4-4fc9-992a-6535ce127a02/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.882671 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869851b9-7ffb-4af0-b166-1d8aa40a5f80" path="/var/lib/kubelet/pods/869851b9-7ffb-4af0-b166-1d8aa40a5f80/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.884339 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7da49224-3b49-41d3-8490-ae2724128e67\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-csf4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-csf4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6jbgs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.885623 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" path="/var/lib/kubelet/pods/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.886365 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92dfbade-90b6-4169-8c07-72cff7f2c82b" path="/var/lib/kubelet/pods/92dfbade-90b6-4169-8c07-72cff7f2c82b/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.892189 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a6e063-3d1a-4d44-875d-185291448c31" path="/var/lib/kubelet/pods/94a6e063-3d1a-4d44-875d-185291448c31/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.893596 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f71a554-e414-4bc3-96d2-674060397afe" path="/var/lib/kubelet/pods/9f71a554-e414-4bc3-96d2-674060397afe/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.900019 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a208c9c2-333b-4b4a-be0d-bc32ec38a821" path="/var/lib/kubelet/pods/a208c9c2-333b-4b4a-be0d-bc32ec38a821/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.903926 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" path="/var/lib/kubelet/pods/a52afe44-fb37-46ed-a1f8-bf39727a3cbe/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.904765 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a555ff2e-0be6-46d5-897d-863bb92ae2b3" path="/var/lib/kubelet/pods/a555ff2e-0be6-46d5-897d-863bb92ae2b3/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.905925 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a88189-c967-4640-879e-27665747f20c" path="/var/lib/kubelet/pods/a7a88189-c967-4640-879e-27665747f20c/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.907169 5110 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="af33e427-6803-48c2-a76a-dd9deb7cbf9a" path="/var/lib/kubelet/pods/af33e427-6803-48c2-a76a-dd9deb7cbf9a/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.907516 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af33e427-6803-48c2-a76a-dd9deb7cbf9a" path="/var/lib/kubelet/pods/af33e427-6803-48c2-a76a-dd9deb7cbf9a/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.911804 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b555f50b-3fcc-4241-b747-d8c898b70530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://6f8f146b9f6264d1b969799f87c68b356d3d701c5836b314a29278bd223e63ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://18527c5fa347f446246a519571a862dc05435325ff8bb4de538c7834fa60f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://638c38a4a2517c3d40ecaeafeb3aff8c788d7679eec37d03bbf83b1d3227cdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e30ed6978ad51bf9888576bae1beb1a9c545283b784e4b0aa1d996642a002616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:42Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://bfcb8305c913d0503cc502dcbfbdedb73b129c8d9b557cf913169eda6efd6d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:41Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://a29edae24756ec647faeaacc3417fe2dc3d9edbd4dd08de37d3d930ea9060a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a29edae24756ec647faeaacc3417fe2dc3d9edbd4dd08de37d3d930ea9060a00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://652b9629cee1d27b3cd8bb6186c5d479f4614b794bd8b255c10e38cc41d4ac0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652b9629cee1d27b3cd8bb6186c5d479f4614b794bd8b255c10e38cc41d4ac0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://65cf9a08bb479461bfc81b04b4184f9881e3354555c9a3b0c77409d162e9b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65cf9a08bb479461bfc81b04b4184f9881e3354555c9a3b0c77409d162e9b15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.913820 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af41de71-79cf-4590-bbe9-9e8b848862cb" path="/var/lib/kubelet/pods/af41de71-79cf-4590-bbe9-9e8b848862cb/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.916105 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" path="/var/lib/kubelet/pods/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.918440 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4750666-1362-4001-abd0-6f89964cc621" path="/var/lib/kubelet/pods/b4750666-1362-4001-abd0-6f89964cc621/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.920284 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b605f283-6f2e-42da-a838-54421690f7d0" path="/var/lib/kubelet/pods/b605f283-6f2e-42da-a838-54421690f7d0/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.920918 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c491984c-7d4b-44aa-8c1e-d7974424fa47" path="/var/lib/kubelet/pods/c491984c-7d4b-44aa-8c1e-d7974424fa47/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.923035 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5f2bfad-70f6-4185-a3d9-81ce12720767" path="/var/lib/kubelet/pods/c5f2bfad-70f6-4185-a3d9-81ce12720767/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.924051 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc85e424-18b2-4924-920b-bd291a8c4b01" path="/var/lib/kubelet/pods/cc85e424-18b2-4924-920b-bd291a8c4b01/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.925276 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce090a97-9ab6-4c40-a719-64ff2acd9778" path="/var/lib/kubelet/pods/ce090a97-9ab6-4c40-a719-64ff2acd9778/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.927978 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d19cb085-0c5b-4810-b654-ce7923221d90" path="/var/lib/kubelet/pods/d19cb085-0c5b-4810-b654-ce7923221d90/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.930509 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" path="/var/lib/kubelet/pods/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.933296 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d565531a-ff86-4608-9d19-767de01ac31b" path="/var/lib/kubelet/pods/d565531a-ff86-4608-9d19-767de01ac31b/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.934190 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e8f42f-dc0e-424b-bb56-5ec849834888" path="/var/lib/kubelet/pods/d7e8f42f-dc0e-424b-bb56-5ec849834888/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.937300 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" path="/var/lib/kubelet/pods/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.938827 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e093be35-bb62-4843-b2e8-094545761610" path="/var/lib/kubelet/pods/e093be35-bb62-4843-b2e8-094545761610/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.939953 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" path="/var/lib/kubelet/pods/e1d2a42d-af1d-4054-9618-ab545e0ed8b7/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.941482 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f559dfa3-3917-43a2-97f6-61ddfda10e93" path="/var/lib/kubelet/pods/f559dfa3-3917-43a2-97f6-61ddfda10e93/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.943546 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65c0ac1-8bca-454d-a2e6-e35cb418beac" path="/var/lib/kubelet/pods/f65c0ac1-8bca-454d-a2e6-e35cb418beac/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.943958 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ce0d9124d558d61880506fe9717799b575efe149b547859b524cd915c8210c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:49:12Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.944742 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" path="/var/lib/kubelet/pods/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.945579 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e2c886-118e-43bb-bef1-c78134de392b" path="/var/lib/kubelet/pods/f7e2c886-118e-43bb-bef1-c78134de392b/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.947204 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" path="/var/lib/kubelet/pods/fc8db2c7-859d-47b3-a900-2bd0c0b2973b/volumes" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.980994 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7ff9391-37cf-48cb-91a1-2ce430fbc515\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://8545c60fc2c493fe31c8c9c36d9eac4f122e0feeea1c08863a4e28302fa86510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://13c01ebb653a6bdaad86850e3714878a54746b6c7cc793fd32d45dade13e7e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13c01ebb653a6bdaad86850e3714878a54746b6c7cc793fd32d45dade13e7e85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.995929 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.995966 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.995978 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.995997 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:12 crc kubenswrapper[5110]: I0317 18:49:12.996009 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:12Z","lastTransitionTime":"2026-03-17T18:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.023961 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331f5282-47de-4f21-903e-5dfb189879f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T18:48:49Z\\\",\\\"message\\\":\\\" envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"ClientsAllowCBOR\\\\\\\" enabled=false\\\\nW0317 18:48:49.126497 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 18:48:49.126663 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI0317 18:48:49.127547 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-97051896/tls.crt::/tmp/serving-cert-97051896/tls.key\\\\\\\"\\\\nI0317 18:48:49.391789 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 18:48:49.393987 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 18:48:49.394005 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 18:48:49.394028 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 18:48:49.394034 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 18:48:49.397362 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 18:48:49.397406 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0317 18:48:49.397405 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0317 18:48:49.397411 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 18:48:49.397432 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 18:48:49.397438 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 18:48:49.397442 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 18:48:49.397445 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0317 18:48:49.399325 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T18:48:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.175224 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.175264 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.175276 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.175291 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.175303 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.178829 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m5g58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baee355c-9055-42f1-87c4-3dc7a08b68f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnc69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnc69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m5g58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.189685 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60643394-bb94-4b51-a9c7-883e2359977a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:47:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e19e709ef87fb9cb61061fe5d588545d4e6145cb814255fd180b74cae14a991f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f91a3deb02a02a3162c98c11c3614f0692f7ccc8366e5de75ac9d079142ac507\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://53b7011b6f8c381cac7dfbb307f74d4a1bc3c33189e417c20d240acaf720cec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T18:47:39Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49d096406a8bfc3c5086a092249671512c15e2309211bd30a18ad8b3c1254237\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:47:37Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:47:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.200550 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.208306 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.220670 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.263045 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-slqfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab807a2a-8744-4ebe-8e74-f841ffbdac5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-slqfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.276525 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.276564 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.276573 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.276587 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.276595 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.314688 5110 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T18:49:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"resources\\\":{},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T18:49:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T18:49:12Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8b652\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T18:49:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4jphj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.430669 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.430715 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.430724 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.430741 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.430751 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.480860 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.481003 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.481043 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.481140 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.481165 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481298 5110 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481365 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:15.481344484 +0000 UTC m=+99.513956026 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481787 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:15.481775856 +0000 UTC m=+99.514387388 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481839 5110 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481879 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:15.481871429 +0000 UTC m=+99.514482961 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481977 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.481991 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.482003 5110 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.482030 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:15.482022603 +0000 UTC m=+99.514634135 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.482101 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.482112 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.482121 5110 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.482182 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:15.482173617 +0000 UTC m=+99.514785149 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.546028 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.546096 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.546108 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.546123 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.546133 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.593569 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.593690 5110 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.594019 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs podName:baee355c-9055-42f1-87c4-3dc7a08b68f6 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:15.593998317 +0000 UTC m=+99.626609839 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs") pod "network-metrics-daemon-m5g58" (UID: "baee355c-9055-42f1-87c4-3dc7a08b68f6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.647506 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" event={"ID":"fe3aacec-3e82-49dc-833e-4b1d848ea2c6","Type":"ContainerStarted","Data":"dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.647735 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" event={"ID":"fe3aacec-3e82-49dc-833e-4b1d848ea2c6","Type":"ContainerStarted","Data":"1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.650107 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"35c38723c7d5b0dbad413c911078841581a267e2686e24599a6df66e026dbea4"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.703165 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.703192 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.703201 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.703213 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.703222 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.708842 5110 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.745247 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.745418 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.745744 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.745795 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.745866 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:13 crc kubenswrapper[5110]: E0317 18:49:13.745913 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.748641 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.748660 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.748668 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.749941 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerStarted","Data":"ba300c7ee193770a83e26680462b04605a6b6a90663e67482b26e32175c2fbda"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.751253 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6k4l7" event={"ID":"d0fce2ad-638a-4ad5-afc3-3cd9c4f125e3","Type":"ContainerStarted","Data":"f8cb3ea63eb55f258aaa3abe01f1a01707bd6d06ae0e3df07a46aa2859b10d6d"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.788287 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-slqfs" event={"ID":"ab807a2a-8744-4ebe-8e74-f841ffbdac5e","Type":"ContainerStarted","Data":"bd42889e0ab53dc7369c158e7b9de2a2fe67172981a2410d51e8de22ef9aa13b"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.855528 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.855769 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.855845 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.855918 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.855982 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.957338 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.957369 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.957378 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.957389 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:13 crc kubenswrapper[5110]: I0317 18:49:13.957399 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:13Z","lastTransitionTime":"2026-03-17T18:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.033290 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" podStartSLOduration=76.033274205 podStartE2EDuration="1m16.033274205s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:13.970580262 +0000 UTC m=+98.003191794" watchObservedRunningTime="2026-03-17 18:49:14.033274205 +0000 UTC m=+98.065885727" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.033656 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=3.033650256 podStartE2EDuration="3.033650256s" podCreationTimestamp="2026-03-17 18:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.030764255 +0000 UTC m=+98.063375797" watchObservedRunningTime="2026-03-17 18:49:14.033650256 +0000 UTC m=+98.066261778" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.085207 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.085270 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.085285 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.085304 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.085315 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.170444 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6kh2q" podStartSLOduration=77.17042483 podStartE2EDuration="1m17.17042483s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.111028798 +0000 UTC m=+98.143640330" watchObservedRunningTime="2026-03-17 18:49:14.17042483 +0000 UTC m=+98.203036342" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.222025 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.222092 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.222102 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.222117 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.222125 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.241854 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.241830476 podStartE2EDuration="3.241830476s" podCreationTimestamp="2026-03-17 18:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.241674092 +0000 UTC m=+98.274285624" watchObservedRunningTime="2026-03-17 18:49:14.241830476 +0000 UTC m=+98.274442018" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.309126 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.309107108 podStartE2EDuration="3.309107108s" podCreationTimestamp="2026-03-17 18:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.308380007 +0000 UTC m=+98.340991569" watchObservedRunningTime="2026-03-17 18:49:14.309107108 +0000 UTC m=+98.341718630" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.323689 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.323737 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.323750 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.323768 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.323781 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.410586 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=3.41056944 podStartE2EDuration="3.41056944s" podCreationTimestamp="2026-03-17 18:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.409886921 +0000 UTC m=+98.442498443" watchObservedRunningTime="2026-03-17 18:49:14.41056944 +0000 UTC m=+98.443180962" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.425637 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.425699 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.425715 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.425736 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.425747 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.528502 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.528671 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.528782 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.528909 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.529024 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.550036 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-slqfs" podStartSLOduration=77.550007519 podStartE2EDuration="1m17.550007519s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.501646053 +0000 UTC m=+98.534257595" watchObservedRunningTime="2026-03-17 18:49:14.550007519 +0000 UTC m=+98.582619071" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.551823 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6k4l7" podStartSLOduration=78.551810719 podStartE2EDuration="1m18.551810719s" podCreationTimestamp="2026-03-17 18:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.550338308 +0000 UTC m=+98.582949830" watchObservedRunningTime="2026-03-17 18:49:14.551810719 +0000 UTC m=+98.584422281" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.685794 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.686086 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.686339 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.686512 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.686683 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.748334 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:14 crc kubenswrapper[5110]: E0317 18:49:14.748938 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.789146 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.789420 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.789430 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.789445 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.789454 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.822837 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.822889 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.828657 5110 generic.go:358] "Generic (PLEG): container finished" podID="341f2a23-d987-42e8-8ca7-f4f596ef3f3d" containerID="ba300c7ee193770a83e26680462b04605a6b6a90663e67482b26e32175c2fbda" exitCode=0 Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.828758 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerDied","Data":"ba300c7ee193770a83e26680462b04605a6b6a90663e67482b26e32175c2fbda"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.913417 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.913473 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.913491 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.913515 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.913534 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:14Z","lastTransitionTime":"2026-03-17T18:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:14 crc kubenswrapper[5110]: I0317 18:49:14.920816 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podStartSLOduration=77.920800642 podStartE2EDuration="1m17.920800642s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:14.644408584 +0000 UTC m=+98.677020106" watchObservedRunningTime="2026-03-17 18:49:14.920800642 +0000 UTC m=+98.953412174" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.089596 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.090018 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.090211 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.090244 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.090263 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.195564 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.195607 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.195616 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.195630 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.195641 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.297802 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.297843 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.297855 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.297872 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.297885 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.400114 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.400150 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.400159 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.400174 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.400185 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.495981 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.496111 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.496138 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.496168 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.496189 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496297 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496312 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496323 5110 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496374 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:19.496358292 +0000 UTC m=+103.528969814 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496694 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:19.496684951 +0000 UTC m=+103.529296473 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496753 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496764 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496771 5110 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496795 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:19.496788594 +0000 UTC m=+103.529400116 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496836 5110 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496860 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:19.496853726 +0000 UTC m=+103.529465248 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496900 5110 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.496921 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:19.496915447 +0000 UTC m=+103.529526969 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.532251 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.532300 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.532310 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.532326 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.532335 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.597209 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.597343 5110 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.597411 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs podName:baee355c-9055-42f1-87c4-3dc7a08b68f6 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:19.597392492 +0000 UTC m=+103.630004024 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs") pod "network-metrics-daemon-m5g58" (UID: "baee355c-9055-42f1-87c4-3dc7a08b68f6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.634404 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.634444 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.634453 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.634470 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.634486 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.736249 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.736293 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.736303 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.736321 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.736333 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.744625 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.744673 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.744781 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.744780 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.745194 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:15 crc kubenswrapper[5110]: E0317 18:49:15.745282 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.885472 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.885525 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.885546 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.885568 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.885585 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.891938 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.894744 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerStarted","Data":"2d1dc3d596d154326f1a657869920b0bd66b394270750f251ab4928beacb8e8e"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.901844 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" event={"ID":"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc","Type":"ContainerStarted","Data":"373edb6ba2c1dc2a4332c406ff7ea1ba69ae19f77bfa7720b9618a035c034b8e"} Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.988457 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.988694 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.988706 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.988721 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:15 crc kubenswrapper[5110]: I0317 18:49:15.988730 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:15Z","lastTransitionTime":"2026-03-17T18:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.090191 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.090224 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.090233 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.090246 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.090255 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.191608 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.191646 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.191655 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.191667 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.191677 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.293304 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.293355 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.293367 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.293386 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.293400 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.396116 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.396164 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.396177 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.396192 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.396203 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.498224 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.498301 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.498321 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.498347 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.498368 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.600967 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.601036 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.601085 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.601119 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.601143 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.697303 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.697360 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.697375 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.697391 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.697405 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.716674 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.716730 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.716745 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.716762 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.716775 5110 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T18:49:16Z","lastTransitionTime":"2026-03-17T18:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.746079 5110 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.746681 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:16 crc kubenswrapper[5110]: E0317 18:49:16.746812 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.749078 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b"] Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.752848 5110 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.757646 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.759632 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\"" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.759851 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"default-dockercfg-hqpm5\"" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.760026 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\"" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.762118 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\"" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.812213 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4a6ece9d-86a4-4db0-a630-6c09172fc298-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.812261 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a6ece9d-86a4-4db0-a630-6c09172fc298-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.812460 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a6ece9d-86a4-4db0-a630-6c09172fc298-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.812517 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4a6ece9d-86a4-4db0-a630-6c09172fc298-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.812592 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4a6ece9d-86a4-4db0-a630-6c09172fc298-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.907292 5110 generic.go:358] "Generic (PLEG): container finished" podID="341f2a23-d987-42e8-8ca7-f4f596ef3f3d" containerID="2d1dc3d596d154326f1a657869920b0bd66b394270750f251ab4928beacb8e8e" exitCode=0 Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.908040 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerDied","Data":"2d1dc3d596d154326f1a657869920b0bd66b394270750f251ab4928beacb8e8e"} Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.912998 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a6ece9d-86a4-4db0-a630-6c09172fc298-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.913034 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4a6ece9d-86a4-4db0-a630-6c09172fc298-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.913077 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4a6ece9d-86a4-4db0-a630-6c09172fc298-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.913124 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4a6ece9d-86a4-4db0-a630-6c09172fc298-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.913147 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a6ece9d-86a4-4db0-a630-6c09172fc298-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.913805 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4a6ece9d-86a4-4db0-a630-6c09172fc298-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.913901 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4a6ece9d-86a4-4db0-a630-6c09172fc298-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.914113 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4a6ece9d-86a4-4db0-a630-6c09172fc298-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.920969 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a6ece9d-86a4-4db0-a630-6c09172fc298-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:16 crc kubenswrapper[5110]: I0317 18:49:16.943769 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a6ece9d-86a4-4db0-a630-6c09172fc298-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-wb66b\" (UID: \"4a6ece9d-86a4-4db0-a630-6c09172fc298\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.073279 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" Mar 17 18:49:17 crc kubenswrapper[5110]: W0317 18:49:17.084637 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a6ece9d_86a4_4db0_a630_6c09172fc298.slice/crio-56f682995b930265ce1c9d74f94e07c92244939a3763a367cb4c449dc8636f6e WatchSource:0}: Error finding container 56f682995b930265ce1c9d74f94e07c92244939a3763a367cb4c449dc8636f6e: Status 404 returned error can't find the container with id 56f682995b930265ce1c9d74f94e07c92244939a3763a367cb4c449dc8636f6e Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.744533 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.744580 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:17 crc kubenswrapper[5110]: E0317 18:49:17.744699 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:17 crc kubenswrapper[5110]: E0317 18:49:17.744861 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.745292 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:17 crc kubenswrapper[5110]: E0317 18:49:17.745577 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.912929 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" event={"ID":"4a6ece9d-86a4-4db0-a630-6c09172fc298","Type":"ContainerStarted","Data":"0b0965f8b906b4a7d873f498b08919b5899e7555a0a9ceaaad8671d0d3502bb9"} Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.913815 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" event={"ID":"4a6ece9d-86a4-4db0-a630-6c09172fc298","Type":"ContainerStarted","Data":"56f682995b930265ce1c9d74f94e07c92244939a3763a367cb4c449dc8636f6e"} Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.923962 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.926519 5110 generic.go:358] "Generic (PLEG): container finished" podID="341f2a23-d987-42e8-8ca7-f4f596ef3f3d" containerID="51b2708e65501a6455085b35539327890d44607d60d28acd7d1c26b300331677" exitCode=0 Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.926607 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerDied","Data":"51b2708e65501a6455085b35539327890d44607d60d28acd7d1c26b300331677"} Mar 17 18:49:17 crc kubenswrapper[5110]: I0317 18:49:17.965102 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-wb66b" podStartSLOduration=80.965045469 podStartE2EDuration="1m20.965045469s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:17.934328805 +0000 UTC m=+101.966940357" watchObservedRunningTime="2026-03-17 18:49:17.965045469 +0000 UTC m=+101.997657011" Mar 17 18:49:18 crc kubenswrapper[5110]: I0317 18:49:18.744296 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:18 crc kubenswrapper[5110]: E0317 18:49:18.744534 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:18 crc kubenswrapper[5110]: I0317 18:49:18.933045 5110 generic.go:358] "Generic (PLEG): container finished" podID="341f2a23-d987-42e8-8ca7-f4f596ef3f3d" containerID="eea0f40ce583b145647a33af19331efecbbbe2b7a4ae2efab389e085cbdf03e8" exitCode=0 Mar 17 18:49:18 crc kubenswrapper[5110]: I0317 18:49:18.933133 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerDied","Data":"eea0f40ce583b145647a33af19331efecbbbe2b7a4ae2efab389e085cbdf03e8"} Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.548748 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.548873 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.548905 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.548945 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.548975 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549135 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549155 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549167 5110 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549224 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:27.549206844 +0000 UTC m=+111.581818366 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549659 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:27.549646846 +0000 UTC m=+111.582258368 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549732 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549745 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549755 5110 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549789 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:27.54977806 +0000 UTC m=+111.582389582 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549843 5110 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549879 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:27.549867552 +0000 UTC m=+111.582479074 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549936 5110 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.549965 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:27.549957235 +0000 UTC m=+111.582568757 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.650434 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.650625 5110 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.650708 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs podName:baee355c-9055-42f1-87c4-3dc7a08b68f6 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:27.650690657 +0000 UTC m=+111.683302179 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs") pod "network-metrics-daemon-m5g58" (UID: "baee355c-9055-42f1-87c4-3dc7a08b68f6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.745444 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.745492 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.745508 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.746119 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.746337 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:19 crc kubenswrapper[5110]: E0317 18:49:19.746475 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.947912 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerStarted","Data":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.949026 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.949075 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.949086 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:19 crc kubenswrapper[5110]: I0317 18:49:19.953973 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerStarted","Data":"dbe71cb904fb858fb4db97024c9dab140b9520d964727351b21a8f1305de2bda"} Mar 17 18:49:20 crc kubenswrapper[5110]: I0317 18:49:20.030912 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:20 crc kubenswrapper[5110]: I0317 18:49:20.030996 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:20 crc kubenswrapper[5110]: I0317 18:49:20.060378 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podStartSLOduration=83.060358242 podStartE2EDuration="1m23.060358242s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:20.03223186 +0000 UTC m=+104.064843382" watchObservedRunningTime="2026-03-17 18:49:20.060358242 +0000 UTC m=+104.092969764" Mar 17 18:49:20 crc kubenswrapper[5110]: I0317 18:49:20.743797 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:20 crc kubenswrapper[5110]: E0317 18:49:20.744306 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:20 crc kubenswrapper[5110]: I0317 18:49:20.963335 5110 generic.go:358] "Generic (PLEG): container finished" podID="341f2a23-d987-42e8-8ca7-f4f596ef3f3d" containerID="dbe71cb904fb858fb4db97024c9dab140b9520d964727351b21a8f1305de2bda" exitCode=0 Mar 17 18:49:20 crc kubenswrapper[5110]: I0317 18:49:20.967249 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerDied","Data":"dbe71cb904fb858fb4db97024c9dab140b9520d964727351b21a8f1305de2bda"} Mar 17 18:49:21 crc kubenswrapper[5110]: I0317 18:49:21.744426 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:21 crc kubenswrapper[5110]: I0317 18:49:21.744544 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:21 crc kubenswrapper[5110]: I0317 18:49:21.744544 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:21 crc kubenswrapper[5110]: E0317 18:49:21.745250 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:21 crc kubenswrapper[5110]: E0317 18:49:21.745262 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:21 crc kubenswrapper[5110]: E0317 18:49:21.745395 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:21 crc kubenswrapper[5110]: I0317 18:49:21.970120 5110 generic.go:358] "Generic (PLEG): container finished" podID="341f2a23-d987-42e8-8ca7-f4f596ef3f3d" containerID="d7630b2885bdfed81680b9adda14a83c38d584fe0e932ffda0e4dd9e1e7dd6c9" exitCode=0 Mar 17 18:49:21 crc kubenswrapper[5110]: I0317 18:49:21.970221 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerDied","Data":"d7630b2885bdfed81680b9adda14a83c38d584fe0e932ffda0e4dd9e1e7dd6c9"} Mar 17 18:49:22 crc kubenswrapper[5110]: I0317 18:49:22.552540 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m5g58"] Mar 17 18:49:22 crc kubenswrapper[5110]: I0317 18:49:22.552691 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:22 crc kubenswrapper[5110]: E0317 18:49:22.552796 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:22 crc kubenswrapper[5110]: I0317 18:49:22.744952 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:22 crc kubenswrapper[5110]: E0317 18:49:22.745336 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:22 crc kubenswrapper[5110]: I0317 18:49:22.745647 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:49:22 crc kubenswrapper[5110]: E0317 18:49:22.745813 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 17 18:49:22 crc kubenswrapper[5110]: I0317 18:49:22.979296 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" event={"ID":"341f2a23-d987-42e8-8ca7-f4f596ef3f3d","Type":"ContainerStarted","Data":"311b306f61653c329c15dbb2ac7fa7805d740d421933dbc4cd94f36003ab8f31"} Mar 17 18:49:23 crc kubenswrapper[5110]: I0317 18:49:23.744507 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:23 crc kubenswrapper[5110]: I0317 18:49:23.744571 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:23 crc kubenswrapper[5110]: E0317 18:49:23.745032 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:23 crc kubenswrapper[5110]: E0317 18:49:23.745132 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:24 crc kubenswrapper[5110]: I0317 18:49:24.744927 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:24 crc kubenswrapper[5110]: E0317 18:49:24.745140 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:24 crc kubenswrapper[5110]: I0317 18:49:24.745669 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:24 crc kubenswrapper[5110]: E0317 18:49:24.745777 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:25 crc kubenswrapper[5110]: I0317 18:49:25.744862 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:25 crc kubenswrapper[5110]: E0317 18:49:25.745148 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 17 18:49:25 crc kubenswrapper[5110]: I0317 18:49:25.745438 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:25 crc kubenswrapper[5110]: E0317 18:49:25.745676 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 17 18:49:26 crc kubenswrapper[5110]: I0317 18:49:26.747637 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:26 crc kubenswrapper[5110]: I0317 18:49:26.747829 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:26 crc kubenswrapper[5110]: E0317 18:49:26.747953 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 17 18:49:26 crc kubenswrapper[5110]: E0317 18:49:26.748232 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m5g58" podUID="baee355c-9055-42f1-87c4-3dc7a08b68f6" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.545668 5110 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeReady" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.545981 5110 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.588979 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.589131 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.589163 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.589197 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.589223 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589353 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589371 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589383 5110 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589441 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:43.589421667 +0000 UTC m=+127.622033199 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589809 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589822 5110 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589831 5110 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589865 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:43.589854959 +0000 UTC m=+127.622466501 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589912 5110 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589940 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:43.589932002 +0000 UTC m=+127.622543534 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589972 5110 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.589995 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:43.589988663 +0000 UTC m=+127.622600195 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.590771 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:43.590742664 +0000 UTC m=+127.623354226 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.609128 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-lh6lx" podStartSLOduration=90.60910762500001 podStartE2EDuration="1m30.609107625s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:23.001045189 +0000 UTC m=+107.033656711" watchObservedRunningTime="2026-03-17 18:49:27.609107625 +0000 UTC m=+111.641719187" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.609873 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.625730 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-lz6ks"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.625897 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.628762 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-mmcpt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.629947 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.633067 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.634432 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-g59zg"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.634692 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.640320 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-xf26c"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.642042 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.642480 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.642656 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"client-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.640493 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.643277 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"client-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.643474 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.643640 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.649821 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.655312 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.655418 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-djmfg\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.643570 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.675014 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.677448 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-client\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.677774 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.677803 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.677914 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.677986 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-qqw4z\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.677936 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.678218 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.678515 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-67c89758df-tnkj2"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.678577 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.678809 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.679268 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-global-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.682923 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"audit-1\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.683304 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-rbac-proxy\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.683459 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-6n5ln\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.683641 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-tls\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.683982 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.684198 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.684629 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.684775 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.684903 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.684993 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-jcmfj\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.685130 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.685226 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-images\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.685663 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.686024 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.688509 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-h787x"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.689095 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690677 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-747b44746d-86ddn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690840 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690866 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ad554df-9a15-43bd-8691-7993f381e08c-images\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690894 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-client-ca\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690917 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c75701-692d-4bb5-a0d5-2020c652ab82-serving-cert\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690932 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-config\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690946 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-tmp\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690969 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690984 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-config\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.690997 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-client-ca\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691022 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/237d41e4-daae-4391-b08d-45b254129ef7-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691038 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-trusted-ca-bundle\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691052 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ad554df-9a15-43bd-8691-7993f381e08c-config\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691114 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pf65\" (UniqueName: \"kubernetes.io/projected/237d41e4-daae-4391-b08d-45b254129ef7-kube-api-access-8pf65\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691136 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-serving-cert\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691156 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/85c75701-692d-4bb5-a0d5-2020c652ab82-tmp\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691171 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691196 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-etcd-client\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691211 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km2h5\" (UniqueName: \"kubernetes.io/projected/6ad554df-9a15-43bd-8691-7993f381e08c-kube-api-access-km2h5\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691252 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-encryption-config\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691267 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ad554df-9a15-43bd-8691-7993f381e08c-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691286 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r5j7\" (UniqueName: \"kubernetes.io/projected/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-kube-api-access-8r5j7\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691315 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/237d41e4-daae-4391-b08d-45b254129ef7-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691329 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-audit-policies\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691343 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b9a2661-d727-4c6b-97cc-7d52281d9425-audit-dir\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691362 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s9r9\" (UniqueName: \"kubernetes.io/projected/85c75701-692d-4bb5-a0d5-2020c652ab82-kube-api-access-4s9r9\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691378 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-serving-cert\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691394 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-etcd-serving-ca\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691408 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t2m5\" (UniqueName: \"kubernetes.io/projected/5b9a2661-d727-4c6b-97cc-7d52281d9425-kube-api-access-6t2m5\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.691426 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237d41e4-daae-4391-b08d-45b254129ef7-config\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.691555 5110 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: E0317 18:49:27.691622 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs podName:baee355c-9055-42f1-87c4-3dc7a08b68f6 nodeName:}" failed. No retries permitted until 2026-03-17 18:49:43.69160453 +0000 UTC m=+127.724216052 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs") pod "network-metrics-daemon-m5g58" (UID: "baee355c-9055-42f1-87c4-3dc7a08b68f6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.702974 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703046 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703415 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703494 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-kl6m8\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703653 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-sjn6s\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703752 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703806 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.703869 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.704313 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.704617 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.704652 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"config-operator-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.704426 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.704483 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.704416 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-wzhvk\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.705304 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-wcbht"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.705929 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.706139 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.708924 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-875ck"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.709196 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.717511 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.717851 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-64d44f6ddf-8jwk2"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.743461 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.743918 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.746354 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.746493 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.749357 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.754764 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"etcd-serving-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.754944 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"encryption-config-1\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.756171 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.756950 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.757489 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.757690 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.760838 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.761770 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.761811 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"audit-1\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.761938 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"image-import-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.762041 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.762428 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.761772 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.765281 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.765457 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.765589 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.767357 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-mdwwj\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.767627 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.767791 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-4zqgh\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.767935 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"etcd-client\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.768332 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.768563 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"dns-operator-dockercfg-wbbsn\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.768691 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"metrics-tls\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.768811 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.768921 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.769040 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.769568 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.770223 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.771048 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8dkm8\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.774303 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-462wb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.774555 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776083 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776200 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776652 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-dockercfg-bf7fj\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776749 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776909 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776984 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777095 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777098 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.776968 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-operator-tls\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777267 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777347 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777366 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"cluster-image-registry-operator-dockercfg-ntnd7\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777567 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-zpddx"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777692 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777974 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.777397 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.778524 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.778611 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-2h6bs\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.778987 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.780166 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.780476 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.780441 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-5hkdz"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.782072 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"service-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.789932 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.790163 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.790440 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.804980 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.805478 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-6tbpn\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.811544 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.811866 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.811991 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-etcd-client\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812023 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812080 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl48g\" (UniqueName: \"kubernetes.io/projected/deb743a6-6308-466e-830a-ae0a66472896-kube-api-access-nl48g\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812101 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-metrics-tls\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812153 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/84f8e2e3-408f-4873-90e9-f778b20e1a32-available-featuregates\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812175 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-etcd-client\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812213 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-km2h5\" (UniqueName: \"kubernetes.io/projected/6ad554df-9a15-43bd-8691-7993f381e08c-kube-api-access-km2h5\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812232 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-image-import-ca\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812249 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-serving-cert\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812267 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/967f9e08-ced9-444e-b03a-439f0f4cc686-config\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812303 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-encryption-config\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812322 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ad554df-9a15-43bd-8691-7993f381e08c-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812339 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-config\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812377 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/c0a5117c-465a-4abf-830d-fac77933f27c-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812404 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8r5j7\" (UniqueName: \"kubernetes.io/projected/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-kube-api-access-8r5j7\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.812530 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.813094 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-config\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.813096 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.813248 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.833556 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"trusted-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834120 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834362 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84f8e2e3-408f-4873-90e9-f778b20e1a32-serving-cert\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834409 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834424 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab45c12d-7f9a-4053-b810-645c966906a1-audit-dir\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834462 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vt82\" (UniqueName: \"kubernetes.io/projected/54c98382-43ad-4113-ab43-b31eb8d4bc58-kube-api-access-8vt82\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834487 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834507 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a5117c-465a-4abf-830d-fac77933f27c-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834563 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-serving-cert\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834578 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-6c46w\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834585 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-oauth-config\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834614 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/237d41e4-daae-4391-b08d-45b254129ef7-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834638 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834662 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ab45c12d-7f9a-4053-b810-645c966906a1-node-pullsecrets\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834696 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834706 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-audit-policies\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834728 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b9a2661-d727-4c6b-97cc-7d52281d9425-audit-dir\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834754 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834779 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-oauth-serving-cert\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834801 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1c85da3-d4ca-4abe-b757-1df684b37eb8-serving-cert\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834823 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-config\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834849 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fngsw\" (UniqueName: \"kubernetes.io/projected/c0a5117c-465a-4abf-830d-fac77933f27c-kube-api-access-fngsw\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834872 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e9bdc21-2fd6-43b9-9666-50637fef0592-trusted-ca\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834902 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4s9r9\" (UniqueName: \"kubernetes.io/projected/85c75701-692d-4bb5-a0d5-2020c652ab82-kube-api-access-4s9r9\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834921 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-serving-cert\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834943 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh4hc\" (UniqueName: \"kubernetes.io/projected/23f9945f-bddd-4c0a-bf2c-cc0baee81974-kube-api-access-xh4hc\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834967 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-etcd-serving-ca\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.834990 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6t2m5\" (UniqueName: \"kubernetes.io/projected/5b9a2661-d727-4c6b-97cc-7d52281d9425-kube-api-access-6t2m5\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835010 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237d41e4-daae-4391-b08d-45b254129ef7-config\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835035 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b5d0e6a-c5e4-409b-a868-61480630c906-config\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835079 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-tmp-dir\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835101 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lg9q\" (UniqueName: \"kubernetes.io/projected/5e9bdc21-2fd6-43b9-9666-50637fef0592-kube-api-access-8lg9q\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835142 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/967f9e08-ced9-444e-b03a-439f0f4cc686-kube-api-access\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835212 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-config\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835236 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835257 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czvqf\" (UniqueName: \"kubernetes.io/projected/ab45c12d-7f9a-4053-b810-645c966906a1-kube-api-access-czvqf\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835293 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/967f9e08-ced9-444e-b03a-439f0f4cc686-serving-cert\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835326 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ad554df-9a15-43bd-8691-7993f381e08c-images\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835358 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88srk\" (UniqueName: \"kubernetes.io/projected/e1c85da3-d4ca-4abe-b757-1df684b37eb8-kube-api-access-88srk\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835386 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-client-ca\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835413 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-service-ca\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835461 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c75701-692d-4bb5-a0d5-2020c652ab82-serving-cert\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.835972 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836120 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/237d41e4-daae-4391-b08d-45b254129ef7-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836180 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9khpm\" (UniqueName: \"kubernetes.io/projected/9731f229-c1b8-4511-819c-b90e6cd6af2e-kube-api-access-9khpm\") pod \"downloads-747b44746d-86ddn\" (UID: \"9731f229-c1b8-4511-819c-b90e6cd6af2e\") " pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836253 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c98382-43ad-4113-ab43-b31eb8d4bc58-serving-cert\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836284 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836352 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-config\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836400 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hst\" (UniqueName: \"kubernetes.io/projected/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-kube-api-access-c4hst\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836429 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6w5m\" (UniqueName: \"kubernetes.io/projected/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-kube-api-access-x6w5m\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836451 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/967f9e08-ced9-444e-b03a-439f0f4cc686-tmp-dir\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836496 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a5117c-465a-4abf-830d-fac77933f27c-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836518 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9bdc21-2fd6-43b9-9666-50637fef0592-config\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836573 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237d41e4-daae-4391-b08d-45b254129ef7-config\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836649 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-client-ca\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.836667 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-audit-policies\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837183 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-tmp\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837216 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-ca\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837236 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b5d0e6a-c5e4-409b-a868-61480630c906-auth-proxy-config\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837256 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0a5117c-465a-4abf-830d-fac77933f27c-tmp\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837290 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/54c98382-43ad-4113-ab43-b31eb8d4bc58-tmp-dir\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837293 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ad554df-9a15-43bd-8691-7993f381e08c-images\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837307 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b5d0e6a-c5e4-409b-a868-61480630c906-machine-approver-tls\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837329 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b9a2661-d727-4c6b-97cc-7d52281d9425-audit-dir\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837324 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-service-ca\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837383 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e9bdc21-2fd6-43b9-9666-50637fef0592-serving-cert\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837430 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-config\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837447 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-client-ca\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837464 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb743a6-6308-466e-830a-ae0a66472896-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837665 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-tmp\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837754 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-config\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.837805 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.838561 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.838730 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839290 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-encryption-config\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839330 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/237d41e4-daae-4391-b08d-45b254129ef7-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839358 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-trusted-ca-bundle\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839378 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0a5117c-465a-4abf-830d-fac77933f27c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839355 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-etcd-serving-ca\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839824 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-trusted-ca-bundle\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839857 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ad554df-9a15-43bd-8691-7993f381e08c-config\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839916 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb743a6-6308-466e-830a-ae0a66472896-config\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.839992 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pckxb\" (UniqueName: \"kubernetes.io/projected/84f8e2e3-408f-4873-90e9-f778b20e1a32-kube-api-access-pckxb\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840115 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-client-ca\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840349 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b9a2661-d727-4c6b-97cc-7d52281d9425-trusted-ca-bundle\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840490 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-audit\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840591 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8pf65\" (UniqueName: \"kubernetes.io/projected/237d41e4-daae-4391-b08d-45b254129ef7-kube-api-access-8pf65\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840642 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-config\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840667 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-client\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840717 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-serving-cert\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840743 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4dxw\" (UniqueName: \"kubernetes.io/projected/6b5d0e6a-c5e4-409b-a868-61480630c906-kube-api-access-f4dxw\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840787 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/85c75701-692d-4bb5-a0d5-2020c652ab82-tmp\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.840816 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-config\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.841322 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/85c75701-692d-4bb5-a0d5-2020c652ab82-tmp\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.841389 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.841673 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ad554df-9a15-43bd-8691-7993f381e08c-config\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.842437 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.844947 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c75701-692d-4bb5-a0d5-2020c652ab82-serving-cert\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.844982 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-serving-cert\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.846097 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ad554df-9a15-43bd-8691-7993f381e08c-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.846317 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-etcd-client\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.846633 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-serving-cert\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.846902 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.847039 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.848383 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/237d41e4-daae-4391-b08d-45b254129ef7-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.849944 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-config\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.852080 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-client\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.852207 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-dockercfg-4vdnc\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.855772 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.855913 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.856614 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5b9a2661-d727-4c6b-97cc-7d52281d9425-encryption-config\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.858856 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.859464 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.866986 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-z7gg7"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.869730 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-service-ca-bundle\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.869759 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-74545575db-lz8zr"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.869995 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.870556 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.872154 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.874773 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.874825 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.874935 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.877189 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.877378 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.879481 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-95rbr"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.879727 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.887968 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.889772 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.890803 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.898206 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-68cf44c8b8-vhklb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.898400 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.906404 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.906626 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.908142 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-session\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.909400 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.909662 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.912234 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.914733 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.915284 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.915502 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.920774 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.921113 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924044 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-875ck"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924077 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-g59zg"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924086 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-lz6ks"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924094 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924102 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d44f6ddf-8jwk2"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924113 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-67c89758df-tnkj2"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924148 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-h787x"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924158 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924165 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924173 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-z7gg7"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924180 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924188 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924196 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-zpddx"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924204 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924203 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924212 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924375 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924402 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.924424 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dhxm4"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.928666 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"openshift-service-ca.crt\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.930663 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-455vz"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.930874 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.934544 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-747b44746d-86ddn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.934597 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jqrnn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.934752 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.938328 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ddn25"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.938475 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941547 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-xf26c"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941569 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-462wb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941584 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941593 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-wcbht"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941601 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941613 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941623 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941634 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941643 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941650 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-74545575db-lz8zr"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941658 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ddn25"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941665 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941674 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-5hkdz"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941685 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941690 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941695 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.941864 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-f8wb2"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942103 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84f8e2e3-408f-4873-90e9-f778b20e1a32-serving-cert\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942128 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab45c12d-7f9a-4053-b810-645c966906a1-audit-dir\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942150 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8vt82\" (UniqueName: \"kubernetes.io/projected/54c98382-43ad-4113-ab43-b31eb8d4bc58-kube-api-access-8vt82\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942171 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942237 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a5117c-465a-4abf-830d-fac77933f27c-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942258 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab45c12d-7f9a-4053-b810-645c966906a1-audit-dir\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942290 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-serving-cert\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942331 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-oauth-config\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942354 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942374 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ab45c12d-7f9a-4053-b810-645c966906a1-node-pullsecrets\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942430 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942451 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-oauth-serving-cert\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942467 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1c85da3-d4ca-4abe-b757-1df684b37eb8-serving-cert\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942518 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-config\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942536 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fngsw\" (UniqueName: \"kubernetes.io/projected/c0a5117c-465a-4abf-830d-fac77933f27c-kube-api-access-fngsw\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942584 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e9bdc21-2fd6-43b9-9666-50637fef0592-trusted-ca\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942606 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xh4hc\" (UniqueName: \"kubernetes.io/projected/23f9945f-bddd-4c0a-bf2c-cc0baee81974-kube-api-access-xh4hc\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942625 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b5d0e6a-c5e4-409b-a868-61480630c906-config\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942642 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-tmp-dir\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942659 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8lg9q\" (UniqueName: \"kubernetes.io/projected/5e9bdc21-2fd6-43b9-9666-50637fef0592-kube-api-access-8lg9q\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942690 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/967f9e08-ced9-444e-b03a-439f0f4cc686-kube-api-access\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942710 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-config\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942725 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942743 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-czvqf\" (UniqueName: \"kubernetes.io/projected/ab45c12d-7f9a-4053-b810-645c966906a1-kube-api-access-czvqf\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942736 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942759 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/967f9e08-ced9-444e-b03a-439f0f4cc686-serving-cert\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942790 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-88srk\" (UniqueName: \"kubernetes.io/projected/e1c85da3-d4ca-4abe-b757-1df684b37eb8-kube-api-access-88srk\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942810 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-service-ca\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942830 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9khpm\" (UniqueName: \"kubernetes.io/projected/9731f229-c1b8-4511-819c-b90e6cd6af2e-kube-api-access-9khpm\") pod \"downloads-747b44746d-86ddn\" (UID: \"9731f229-c1b8-4511-819c-b90e6cd6af2e\") " pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942845 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c98382-43ad-4113-ab43-b31eb8d4bc58-serving-cert\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942863 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942882 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hst\" (UniqueName: \"kubernetes.io/projected/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-kube-api-access-c4hst\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942900 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x6w5m\" (UniqueName: \"kubernetes.io/projected/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-kube-api-access-x6w5m\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942918 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/967f9e08-ced9-444e-b03a-439f0f4cc686-tmp-dir\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942936 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a5117c-465a-4abf-830d-fac77933f27c-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942954 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9bdc21-2fd6-43b9-9666-50637fef0592-config\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942975 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-ca\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.942992 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b5d0e6a-c5e4-409b-a868-61480630c906-auth-proxy-config\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943009 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0a5117c-465a-4abf-830d-fac77933f27c-tmp\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943030 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/54c98382-43ad-4113-ab43-b31eb8d4bc58-tmp-dir\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943046 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b5d0e6a-c5e4-409b-a868-61480630c906-machine-approver-tls\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943080 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-service-ca\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943095 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e9bdc21-2fd6-43b9-9666-50637fef0592-serving-cert\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943135 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8606fbd-7611-4970-9f65-f3e12bb76894-webhook-certs\") pod \"multus-admission-controller-69db94689b-z7gg7\" (UID: \"f8606fbd-7611-4970-9f65-f3e12bb76894\") " pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943165 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb743a6-6308-466e-830a-ae0a66472896-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943199 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-encryption-config\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943215 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-trusted-ca-bundle\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943233 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0a5117c-465a-4abf-830d-fac77933f27c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943254 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb743a6-6308-466e-830a-ae0a66472896-config\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943270 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pckxb\" (UniqueName: \"kubernetes.io/projected/84f8e2e3-408f-4873-90e9-f778b20e1a32-kube-api-access-pckxb\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943289 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-audit\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943307 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-config\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943324 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-client\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943344 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f4dxw\" (UniqueName: \"kubernetes.io/projected/6b5d0e6a-c5e4-409b-a868-61480630c906-kube-api-access-f4dxw\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943361 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfv6f\" (UniqueName: \"kubernetes.io/projected/f8606fbd-7611-4970-9f65-f3e12bb76894-kube-api-access-tfv6f\") pod \"multus-admission-controller-69db94689b-z7gg7\" (UID: \"f8606fbd-7611-4970-9f65-f3e12bb76894\") " pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943683 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/967f9e08-ced9-444e-b03a-439f0f4cc686-tmp-dir\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943725 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-config\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943754 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-etcd-client\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943792 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nl48g\" (UniqueName: \"kubernetes.io/projected/deb743a6-6308-466e-830a-ae0a66472896-kube-api-access-nl48g\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943810 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-metrics-tls\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943834 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/84f8e2e3-408f-4873-90e9-f778b20e1a32-available-featuregates\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943857 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-image-import-ca\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943873 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-serving-cert\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943888 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/967f9e08-ced9-444e-b03a-439f0f4cc686-config\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943908 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-config\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.943923 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/c0a5117c-465a-4abf-830d-fac77933f27c-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944096 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-service-ca\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944360 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/c0a5117c-465a-4abf-830d-fac77933f27c-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944459 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-service-ca\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944493 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-ca\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944810 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb743a6-6308-466e-830a-ae0a66472896-config\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944833 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-audit\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.944915 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b5d0e6a-c5e4-409b-a868-61480630c906-auth-proxy-config\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.945183 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0a5117c-465a-4abf-830d-fac77933f27c-tmp\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.945318 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c98382-43ad-4113-ab43-b31eb8d4bc58-config\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.945417 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/54c98382-43ad-4113-ab43-b31eb8d4bc58-tmp-dir\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.945598 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a5117c-465a-4abf-830d-fac77933f27c-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.945870 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-serving-cert\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.945954 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-config\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.946433 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9bdc21-2fd6-43b9-9666-50637fef0592-config\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.946669 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.947161 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/967f9e08-ced9-444e-b03a-439f0f4cc686-serving-cert\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.947290 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b5d0e6a-c5e4-409b-a868-61480630c906-config\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.947571 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c98382-43ad-4113-ab43-b31eb8d4bc58-serving-cert\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.947637 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-tmp-dir\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.947766 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e9bdc21-2fd6-43b9-9666-50637fef0592-trusted-ca\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.948367 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-config\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.948940 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-encryption-config\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.949301 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ab45c12d-7f9a-4053-b810-645c966906a1-node-pullsecrets\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.949448 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-config\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.949704 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/967f9e08-ced9-444e-b03a-439f0f4cc686-config\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.949957 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950188 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84f8e2e3-408f-4873-90e9-f778b20e1a32-serving-cert\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950236 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1c85da3-d4ca-4abe-b757-1df684b37eb8-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950348 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\"" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950486 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/84f8e2e3-408f-4873-90e9-f778b20e1a32-available-featuregates\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950606 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-trusted-ca-bundle\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950735 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-image-import-ca\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950747 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/23f9945f-bddd-4c0a-bf2c-cc0baee81974-oauth-serving-cert\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.950903 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab45c12d-7f9a-4053-b810-645c966906a1-config\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.951506 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-etcd-client\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.951522 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/23f9945f-bddd-4c0a-bf2c-cc0baee81974-console-oauth-config\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.952018 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b5d0e6a-c5e4-409b-a868-61480630c906-machine-approver-tls\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.952402 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54c98382-43ad-4113-ab43-b31eb8d4bc58-etcd-client\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.952766 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e9bdc21-2fd6-43b9-9666-50637fef0592-serving-cert\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.952848 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb743a6-6308-466e-830a-ae0a66472896-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.953078 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0a5117c-465a-4abf-830d-fac77933f27c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.954043 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-metrics-tls\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.954631 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab45c12d-7f9a-4053-b810-645c966906a1-serving-cert\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.954695 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1c85da3-d4ca-4abe-b757-1df684b37eb8-serving-cert\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.957660 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958427 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dhxm4"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958459 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-f8wb2"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958473 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958484 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958496 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958508 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-95rbr"] Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.958627 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:27 crc kubenswrapper[5110]: I0317 18:49:27.968689 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.001266 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.008625 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.028791 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.044882 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8606fbd-7611-4970-9f65-f3e12bb76894-webhook-certs\") pod \"multus-admission-controller-69db94689b-z7gg7\" (UID: \"f8606fbd-7611-4970-9f65-f3e12bb76894\") " pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.044978 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tfv6f\" (UniqueName: \"kubernetes.io/projected/f8606fbd-7611-4970-9f65-f3e12bb76894-kube-api-access-tfv6f\") pod \"multus-admission-controller-69db94689b-z7gg7\" (UID: \"f8606fbd-7611-4970-9f65-f3e12bb76894\") " pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.048396 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.068947 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-d2bf2\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.088018 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.108553 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"audit\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.129125 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.148290 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.177096 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.188622 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"kube-root-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.256718 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r5j7\" (UniqueName: \"kubernetes.io/projected/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-kube-api-access-8r5j7\") pod \"route-controller-manager-776cdc94d6-m9q7t\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.269687 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-km2h5\" (UniqueName: \"kubernetes.io/projected/6ad554df-9a15-43bd-8691-7993f381e08c-kube-api-access-km2h5\") pod \"machine-api-operator-755bb95488-xf26c\" (UID: \"6ad554df-9a15-43bd-8691-7993f381e08c\") " pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.286792 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s9r9\" (UniqueName: \"kubernetes.io/projected/85c75701-692d-4bb5-a0d5-2020c652ab82-kube-api-access-4s9r9\") pod \"controller-manager-65b6cccf98-lz6ks\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.307455 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.308166 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.317012 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t2m5\" (UniqueName: \"kubernetes.io/projected/5b9a2661-d727-4c6b-97cc-7d52281d9425-kube-api-access-6t2m5\") pod \"apiserver-8596bd845d-g59zg\" (UID: \"5b9a2661-d727-4c6b-97cc-7d52281d9425\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.328601 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.348305 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.368285 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.405954 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pf65\" (UniqueName: \"kubernetes.io/projected/237d41e4-daae-4391-b08d-45b254129ef7-kube-api-access-8pf65\") pod \"openshift-controller-manager-operator-686468bdd5-vnjj4\" (UID: \"237d41e4-daae-4391-b08d-45b254129ef7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.408586 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"pprof-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.428782 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"openshift-service-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.449198 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"kube-root-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.467969 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.489342 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serviceaccount-dockercfg-4gqzj\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.508455 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"package-server-manager-serving-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.528657 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-dockercfg-2wbn2\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.538139 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-xf26c"] Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.543590 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:28 crc kubenswrapper[5110]: W0317 18:49:28.546470 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ad554df_9a15_43bd_8691_7993f381e08c.slice/crio-c0cfcb241f2d0a5dd669522efe77cb56a5b7a12495525aea575c48152b822544 WatchSource:0}: Error finding container c0cfcb241f2d0a5dd669522efe77cb56a5b7a12495525aea575c48152b822544: Status 404 returned error can't find the container with id c0cfcb241f2d0a5dd669522efe77cb56a5b7a12495525aea575c48152b822544 Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.548661 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-scheduler-operator-serving-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.568469 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-config\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.578679 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.589214 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.591849 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.610094 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serving-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.627128 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.628723 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"catalog-operator-serving-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.649188 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-admission-controller-secret\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.659910 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8606fbd-7611-4970-9f65-f3e12bb76894-webhook-certs\") pod \"multus-admission-controller-69db94689b-z7gg7\" (UID: \"f8606fbd-7611-4970-9f65-f3e12bb76894\") " pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.669938 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ac-dockercfg-gj7jx\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.689920 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bgxvm\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.709200 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.729130 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.748486 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.768941 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.769273 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.769963 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.788028 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"packageserver-service-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.808265 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.838282 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.840995 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t"] Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.848100 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-bjqfd\"" Mar 17 18:49:28 crc kubenswrapper[5110]: W0317 18:49:28.851480 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc6f0b62_c129_4c6d_a228_af2e48b6a0ce.slice/crio-92da9221935705fd7f5adc4821852df2a489349d90da00979af52c15cd24de13 WatchSource:0}: Error finding container 92da9221935705fd7f5adc4821852df2a489349d90da00979af52c15cd24de13: Status 404 returned error can't find the container with id 92da9221935705fd7f5adc4821852df2a489349d90da00979af52c15cd24de13 Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.868607 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.881860 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4"] Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.882159 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-g59zg"] Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.886802 5110 request.go:752] "Waited before sending request" delay="1.009095448s" reason="client-side throttling, not priority and fairness" verb="GET" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.888301 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.907887 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.929074 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-jmhxf\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.948579 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:28 crc kubenswrapper[5110]: I0317 18:49:28.970459 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.007347 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" event={"ID":"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce","Type":"ContainerStarted","Data":"bf7af23e9651c1dc674d6207c192895da2ffeee1ce358fb60dfdda02626e311d"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.007401 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" event={"ID":"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce","Type":"ContainerStarted","Data":"92da9221935705fd7f5adc4821852df2a489349d90da00979af52c15cd24de13"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.008300 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.008584 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" event={"ID":"237d41e4-daae-4391-b08d-45b254129ef7","Type":"ContainerStarted","Data":"cc4dc22aadc9d194ffe9101b9f91b1deeffbd33687b4f04ef0689a45737c310a"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.009226 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" event={"ID":"5b9a2661-d727-4c6b-97cc-7d52281d9425","Type":"ContainerStarted","Data":"5c9a25a0e21e4ea8bda634f518badb09bc365f343449d1403eaa4aef552cab7a"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.011487 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" event={"ID":"6ad554df-9a15-43bd-8691-7993f381e08c","Type":"ContainerStarted","Data":"1a02a6d05a0e64bea45f2ccc16165cfb589c1e92016cd7e9345032b372d9ceba"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.011513 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" event={"ID":"6ad554df-9a15-43bd-8691-7993f381e08c","Type":"ContainerStarted","Data":"019831800b1d2dc5053e615141dd31dfa00fc1b446950caea0b7d484c9acaf41"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.011524 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" event={"ID":"6ad554df-9a15-43bd-8691-7993f381e08c","Type":"ContainerStarted","Data":"c0cfcb241f2d0a5dd669522efe77cb56a5b7a12495525aea575c48152b822544"} Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.028293 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-dockercfg-2cfkp\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.052616 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-metrics\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.074914 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-lz6ks"] Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.076954 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"marketplace-trusted-ca\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.088525 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.108552 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"ingress-operator-dockercfg-74nwh\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.128445 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"metrics-tls\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.154580 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"trusted-ca\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.168459 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.188707 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.208221 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.228861 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-certs-default\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.249609 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.269254 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.289014 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-kw8fx\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.308847 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.337178 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.349813 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.368486 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-dockercfg-tnfx9\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.388495 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-serving-cert\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.409012 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-config\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.429801 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-tls\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.448038 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-dockercfg-gnx66\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.469091 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-controller-dockercfg-xnj77\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.488916 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mcc-proxy-tls\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.509403 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-dockercfg-sw6nc\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.531489 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-images\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.549399 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mco-proxy-tls\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.570668 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-kknhg\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.588675 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.608786 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.630032 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-kpvmz\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.649100 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.668871 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.688720 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-dockercfg-dzw6b\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.709337 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-tls\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.729196 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"node-bootstrapper-token\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.749085 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-sysctl-allowlist\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.768619 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"hostpath-provisioner\"/\"csi-hostpath-provisioner-sa-dockercfg-7dcws\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.788791 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"openshift-service-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.808739 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"kube-root-ca.crt\"" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.849151 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vt82\" (UniqueName: \"kubernetes.io/projected/54c98382-43ad-4113-ab43-b31eb8d4bc58-kube-api-access-8vt82\") pod \"etcd-operator-69b85846b6-462wb\" (UID: \"54c98382-43ad-4113-ab43-b31eb8d4bc58\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.877072 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hst\" (UniqueName: \"kubernetes.io/projected/c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196-kube-api-access-c4hst\") pod \"kube-storage-version-migrator-operator-565b79b866-ltfln\" (UID: \"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.882900 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a5117c-465a-4abf-830d-fac77933f27c-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.905684 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-88srk\" (UniqueName: \"kubernetes.io/projected/e1c85da3-d4ca-4abe-b757-1df684b37eb8-kube-api-access-88srk\") pod \"authentication-operator-7f5c659b84-bnwct\" (UID: \"e1c85da3-d4ca-4abe-b757-1df684b37eb8\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.907392 5110 request.go:752] "Waited before sending request" delay="1.963989149s" reason="client-side throttling, not priority and fairness" verb="POST" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/serviceaccounts/dns-operator/token" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.930023 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6w5m\" (UniqueName: \"kubernetes.io/projected/1bda76e2-be33-4ca9-b962-c64fa6d4df1e-kube-api-access-x6w5m\") pod \"dns-operator-799b87ffcd-875ck\" (UID: \"1bda76e2-be33-4ca9-b962-c64fa6d4df1e\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.942643 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pckxb\" (UniqueName: \"kubernetes.io/projected/84f8e2e3-408f-4873-90e9-f778b20e1a32-kube-api-access-pckxb\") pod \"openshift-config-operator-5777786469-h787x\" (UID: \"84f8e2e3-408f-4873-90e9-f778b20e1a32\") " pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.965547 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9khpm\" (UniqueName: \"kubernetes.io/projected/9731f229-c1b8-4511-819c-b90e6cd6af2e-kube-api-access-9khpm\") pod \"downloads-747b44746d-86ddn\" (UID: \"9731f229-c1b8-4511-819c-b90e6cd6af2e\") " pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.981738 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" Mar 17 18:49:29 crc kubenswrapper[5110]: I0317 18:49:29.985153 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lg9q\" (UniqueName: \"kubernetes.io/projected/5e9bdc21-2fd6-43b9-9666-50637fef0592-kube-api-access-8lg9q\") pod \"console-operator-67c89758df-tnkj2\" (UID: \"5e9bdc21-2fd6-43b9-9666-50637fef0592\") " pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.008185 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh4hc\" (UniqueName: \"kubernetes.io/projected/23f9945f-bddd-4c0a-bf2c-cc0baee81974-kube-api-access-xh4hc\") pod \"console-64d44f6ddf-8jwk2\" (UID: \"23f9945f-bddd-4c0a-bf2c-cc0baee81974\") " pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.010648 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.017941 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" event={"ID":"237d41e4-daae-4391-b08d-45b254129ef7","Type":"ContainerStarted","Data":"6f439eaa909d81271e5a8d8f08dd0ca5717bd6fe0a77bfb755dcaf5fd9ca32fd"} Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.019456 5110 generic.go:358] "Generic (PLEG): container finished" podID="5b9a2661-d727-4c6b-97cc-7d52281d9425" containerID="d753007c9eb5ff91f0864cdaad5c28bd0a1ff18a2a6e9cb316c593a9429503e4" exitCode=0 Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.019574 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" event={"ID":"5b9a2661-d727-4c6b-97cc-7d52281d9425","Type":"ContainerDied","Data":"d753007c9eb5ff91f0864cdaad5c28bd0a1ff18a2a6e9cb316c593a9429503e4"} Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.021159 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" event={"ID":"85c75701-692d-4bb5-a0d5-2020c652ab82","Type":"ContainerStarted","Data":"54a40f6fcadf19fd5733bfcd79d1e67f1f317709bf00d13f8d23ad9bd8bc8df8"} Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.021213 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" event={"ID":"85c75701-692d-4bb5-a0d5-2020c652ab82","Type":"ContainerStarted","Data":"797be210166a85a6cd44dfe58626d57eaf7f4c11cee1440db3f51fb2cf8696fa"} Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.021432 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.021460 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.026221 5110 patch_prober.go:28] interesting pod/controller-manager-65b6cccf98-lz6ks container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.026275 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.027096 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/967f9e08-ced9-444e-b03a-439f0f4cc686-kube-api-access\") pod \"kube-apiserver-operator-575994946d-k5fgh\" (UID: \"967f9e08-ced9-444e-b03a-439f0f4cc686\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.039419 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.046776 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-czvqf\" (UniqueName: \"kubernetes.io/projected/ab45c12d-7f9a-4053-b810-645c966906a1-kube-api-access-czvqf\") pod \"apiserver-9ddfb9f55-wcbht\" (UID: \"ab45c12d-7f9a-4053-b810-645c966906a1\") " pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.074487 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4dxw\" (UniqueName: \"kubernetes.io/projected/6b5d0e6a-c5e4-409b-a868-61480630c906-kube-api-access-f4dxw\") pod \"machine-approver-54c688565-hhs4z\" (UID: \"6b5d0e6a-c5e4-409b-a868-61480630c906\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.082819 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.085972 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl48g\" (UniqueName: \"kubernetes.io/projected/deb743a6-6308-466e-830a-ae0a66472896-kube-api-access-nl48g\") pod \"openshift-apiserver-operator-846cbfc458-7t4sr\" (UID: \"deb743a6-6308-466e-830a-ae0a66472896\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.107216 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fngsw\" (UniqueName: \"kubernetes.io/projected/c0a5117c-465a-4abf-830d-fac77933f27c-kube-api-access-fngsw\") pod \"cluster-image-registry-operator-86c45576b9-f8xnn\" (UID: \"c0a5117c-465a-4abf-830d-fac77933f27c\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.116158 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.129382 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9pgs7\"" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.139013 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.148776 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.158327 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.168451 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.169198 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.214860 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfv6f\" (UniqueName: \"kubernetes.io/projected/f8606fbd-7611-4970-9f65-f3e12bb76894-kube-api-access-tfv6f\") pod \"multus-admission-controller-69db94689b-z7gg7\" (UID: \"f8606fbd-7611-4970-9f65-f3e12bb76894\") " pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.216775 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-875ck"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.229206 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.238379 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.248593 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.273547 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287230 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7930ff4e-b5eb-42d7-ab81-884f18b423d7-config-volume\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287271 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/33efc255-a7a4-4abf-93a8-a522763c8141-tmpfs\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287289 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287306 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-dir\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287321 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287347 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152126f1-6413-47f3-9847-ca6e3e561c7b-config\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287363 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287382 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287397 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287422 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287438 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-installation-pull-secrets\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287454 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fb509cc-d224-4492-8b34-b7625127411a-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287467 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69e1d569-06c8-4098-8e3f-57c1a5970756-apiservice-cert\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287489 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-tls\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287507 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287559 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-certificates\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287589 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fb509cc-d224-4492-8b34-b7625127411a-srv-cert\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287615 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-tmp\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287641 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8w7l\" (UniqueName: \"kubernetes.io/projected/33efc255-a7a4-4abf-93a8-a522763c8141-kube-api-access-k8w7l\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287656 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-bound-sa-token\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287673 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287702 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/152126f1-6413-47f3-9847-ca6e3e561c7b-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287728 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69e1d569-06c8-4098-8e3f-57c1a5970756-webhook-cert\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287746 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287781 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7930ff4e-b5eb-42d7-ab81-884f18b423d7-secret-volume\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287797 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/33efc255-a7a4-4abf-93a8-a522763c8141-srv-cert\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287817 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287830 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69e1d569-06c8-4098-8e3f-57c1a5970756-tmpfs\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287858 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gvc\" (UniqueName: \"kubernetes.io/projected/8477e2b6-663f-49b8-a714-5925de307ac9-kube-api-access-57gvc\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287887 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srk6n\" (UniqueName: \"kubernetes.io/projected/7930ff4e-b5eb-42d7-ab81-884f18b423d7-kube-api-access-srk6n\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287931 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.287947 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2337485-1832-41b3-97d8-4ccd31ed60a3-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-gg6pg\" (UID: \"e2337485-1832-41b3-97d8-4ccd31ed60a3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.288070 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8477e2b6-663f-49b8-a714-5925de307ac9-signing-key\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.288089 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4tsw\" (UniqueName: \"kubernetes.io/projected/5fb509cc-d224-4492-8b34-b7625127411a-kube-api-access-z4tsw\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.288108 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/33efc255-a7a4-4abf-93a8-a522763c8141-profile-collector-cert\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.288125 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/152126f1-6413-47f3-9847-ca6e3e561c7b-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.288265 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cmlt\" (UniqueName: \"kubernetes.io/projected/69e1d569-06c8-4098-8e3f-57c1a5970756-kube-api-access-9cmlt\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.289815 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c5ac55-d0d1-4075-84c6-3de5c8388a96-config\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.289863 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vml5s\" (UniqueName: \"kubernetes.io/projected/97c5ac55-d0d1-4075-84c6-3de5c8388a96-kube-api-access-vml5s\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.289920 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8477e2b6-663f-49b8-a714-5925de307ac9-signing-cabundle\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.290402 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/152126f1-6413-47f3-9847-ca6e3e561c7b-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.290478 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnpvm\" (UniqueName: \"kubernetes.io/projected/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-kube-api-access-nnpvm\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291146 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwzw\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-kube-api-access-4nwzw\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291217 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-policies\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291237 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5tdt\" (UniqueName: \"kubernetes.io/projected/e2337485-1832-41b3-97d8-4ccd31ed60a3-kube-api-access-n5tdt\") pod \"package-server-manager-77f986bd66-gg6pg\" (UID: \"e2337485-1832-41b3-97d8-4ccd31ed60a3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.291292 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:30.791273361 +0000 UTC m=+114.823884883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291336 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291347 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291578 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-trusted-ca\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291608 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5fb509cc-d224-4492-8b34-b7625127411a-tmpfs\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.291925 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-ca-trust-extracted\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.292168 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.292201 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.292221 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm7qv\" (UniqueName: \"kubernetes.io/projected/c2d80dcc-dd1a-434c-885d-16a7db229591-kube-api-access-vm7qv\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.292471 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgqwt\" (UniqueName: \"kubernetes.io/projected/eb2106ad-091a-45e1-8e42-7dd01f174c03-kube-api-access-wgqwt\") pod \"cluster-samples-operator-6b564684c8-jx99w\" (UID: \"eb2106ad-091a-45e1-8e42-7dd01f174c03\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.292545 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97c5ac55-d0d1-4075-84c6-3de5c8388a96-serving-cert\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.292568 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb2106ad-091a-45e1-8e42-7dd01f174c03-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-jx99w\" (UID: \"eb2106ad-091a-45e1-8e42-7dd01f174c03\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.302939 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.319327 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.353000 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.364597 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.400389 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.400587 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/88ddb708-6eb5-4d13-ab7c-798b0a120818-certs\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.400624 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-policies\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.400648 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n5tdt\" (UniqueName: \"kubernetes.io/projected/e2337485-1832-41b3-97d8-4ccd31ed60a3-kube-api-access-n5tdt\") pod \"package-server-manager-77f986bd66-gg6pg\" (UID: \"e2337485-1832-41b3-97d8-4ccd31ed60a3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.400674 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.400718 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-trusted-ca\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.401137 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5fb509cc-d224-4492-8b34-b7625127411a-tmpfs\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.402021 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-policies\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.402146 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5fb509cc-d224-4492-8b34-b7625127411a-tmpfs\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.402249 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:30.902153655 +0000 UTC m=+114.934765177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.402876 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403250 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-ca-trust-extracted\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403349 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403418 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403489 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vm7qv\" (UniqueName: \"kubernetes.io/projected/c2d80dcc-dd1a-434c-885d-16a7db229591-kube-api-access-vm7qv\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403562 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wgqwt\" (UniqueName: \"kubernetes.io/projected/eb2106ad-091a-45e1-8e42-7dd01f174c03-kube-api-access-wgqwt\") pod \"cluster-samples-operator-6b564684c8-jx99w\" (UID: \"eb2106ad-091a-45e1-8e42-7dd01f174c03\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403636 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97c5ac55-d0d1-4075-84c6-3de5c8388a96-serving-cert\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403702 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb2106ad-091a-45e1-8e42-7dd01f174c03-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-jx99w\" (UID: \"eb2106ad-091a-45e1-8e42-7dd01f174c03\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403791 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bx9q\" (UniqueName: \"kubernetes.io/projected/88ddb708-6eb5-4d13-ab7c-798b0a120818-kube-api-access-5bx9q\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403857 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7930ff4e-b5eb-42d7-ab81-884f18b423d7-config-volume\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.403943 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/33efc255-a7a4-4abf-93a8-a522763c8141-tmpfs\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404006 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404097 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-trusted-ca\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404170 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-dir\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404413 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404478 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152126f1-6413-47f3-9847-ca6e3e561c7b-config\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404543 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-registration-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404610 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404678 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-mountpoint-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404749 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404811 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404876 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-stats-auth\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.404944 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mrk8\" (UniqueName: \"kubernetes.io/projected/e599baf2-11e5-4d78-9282-b6850af2bd7e-kube-api-access-7mrk8\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.405015 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.406582 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74bf2564-afaa-4967-9901-3fd670b5b991-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.406868 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74bf2564-afaa-4967-9901-3fd670b5b991-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.406947 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-installation-pull-secrets\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407024 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fb509cc-d224-4492-8b34-b7625127411a-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407103 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69e1d569-06c8-4098-8e3f-57c1a5970756-apiservice-cert\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407179 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w2pz\" (UniqueName: \"kubernetes.io/projected/f7edb507-fedd-4535-b7fe-46095dbd1b5c-kube-api-access-9w2pz\") pod \"ingress-canary-f8wb2\" (UID: \"f7edb507-fedd-4535-b7fe-46095dbd1b5c\") " pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407250 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-tls\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407329 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407396 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-socket-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407458 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6kmt\" (UniqueName: \"kubernetes.io/projected/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-kube-api-access-n6kmt\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407541 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74bf2564-afaa-4967-9901-3fd670b5b991-config\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407629 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/023fe200-90c4-43c0-aa37-2241d9061e06-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407715 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-certificates\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407794 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fb509cc-d224-4492-8b34-b7625127411a-srv-cert\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407859 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/57ae9010-5764-463a-93ad-d4036ee2b53c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-xsmjk\" (UID: \"57ae9010-5764-463a-93ad-d4036ee2b53c\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.407954 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-tmp\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.408026 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-k8w7l\" (UniqueName: \"kubernetes.io/projected/33efc255-a7a4-4abf-93a8-a522763c8141-kube-api-access-k8w7l\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.408585 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.410302 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411215 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs7rs\" (UniqueName: \"kubernetes.io/projected/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-kube-api-access-rs7rs\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411287 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411360 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/10e07288-52f4-407c-a8eb-709214e358f5-ready\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411433 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-bound-sa-token\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411521 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411592 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-images\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411655 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10e07288-52f4-407c-a8eb-709214e358f5-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411726 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-default-certificate\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411835 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/152126f1-6413-47f3-9847-ca6e3e561c7b-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.411927 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69e1d569-06c8-4098-8e3f-57c1a5970756-webhook-cert\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412016 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412139 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b68f5\" (UniqueName: \"kubernetes.io/projected/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-kube-api-access-b68f5\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412227 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7930ff4e-b5eb-42d7-ab81-884f18b423d7-secret-volume\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412299 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/33efc255-a7a4-4abf-93a8-a522763c8141-srv-cert\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412366 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412444 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69e1d569-06c8-4098-8e3f-57c1a5970756-tmpfs\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412530 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412601 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-metrics-certs\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412788 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152126f1-6413-47f3-9847-ca6e3e561c7b-config\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412937 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/023fe200-90c4-43c0-aa37-2241d9061e06-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.412996 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-57gvc\" (UniqueName: \"kubernetes.io/projected/8477e2b6-663f-49b8-a714-5925de307ac9-kube-api-access-57gvc\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413027 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-plugins-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413044 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mb6c\" (UniqueName: \"kubernetes.io/projected/023fe200-90c4-43c0-aa37-2241d9061e06-kube-api-access-4mb6c\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413081 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-srk6n\" (UniqueName: \"kubernetes.io/projected/7930ff4e-b5eb-42d7-ab81-884f18b423d7-kube-api-access-srk6n\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413103 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpds5\" (UniqueName: \"kubernetes.io/projected/7923cd83-a73e-4a4b-9bdd-b38f7656579b-kube-api-access-fpds5\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413112 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/33efc255-a7a4-4abf-93a8-a522763c8141-tmpfs\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413119 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/88ddb708-6eb5-4d13-ab7c-798b0a120818-node-bootstrap-token\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413198 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413221 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2337485-1832-41b3-97d8-4ccd31ed60a3-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-gg6pg\" (UID: \"e2337485-1832-41b3-97d8-4ccd31ed60a3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413274 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413294 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f7edb507-fedd-4535-b7fe-46095dbd1b5c-cert\") pod \"ingress-canary-f8wb2\" (UID: \"f7edb507-fedd-4535-b7fe-46095dbd1b5c\") " pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413310 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7923cd83-a73e-4a4b-9bdd-b38f7656579b-tmp-dir\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413334 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413373 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jm87\" (UniqueName: \"kubernetes.io/projected/10e07288-52f4-407c-a8eb-709214e358f5-kube-api-access-2jm87\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413390 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kpbh\" (UniqueName: \"kubernetes.io/projected/9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd-kube-api-access-5kpbh\") pod \"migrator-866fcbc849-xkvs8\" (UID: \"9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413409 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht2w7\" (UniqueName: \"kubernetes.io/projected/57ae9010-5764-463a-93ad-d4036ee2b53c-kube-api-access-ht2w7\") pod \"control-plane-machine-set-operator-75ffdb6fcd-xsmjk\" (UID: \"57ae9010-5764-463a-93ad-d4036ee2b53c\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413437 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8477e2b6-663f-49b8-a714-5925de307ac9-signing-key\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413456 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-z4tsw\" (UniqueName: \"kubernetes.io/projected/5fb509cc-d224-4492-8b34-b7625127411a-kube-api-access-z4tsw\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413477 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/33efc255-a7a4-4abf-93a8-a522763c8141-profile-collector-cert\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413494 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/152126f1-6413-47f3-9847-ca6e3e561c7b-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413511 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9cmlt\" (UniqueName: \"kubernetes.io/projected/69e1d569-06c8-4098-8e3f-57c1a5970756-kube-api-access-9cmlt\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413548 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c5ac55-d0d1-4075-84c6-3de5c8388a96-config\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413575 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vml5s\" (UniqueName: \"kubernetes.io/projected/97c5ac55-d0d1-4075-84c6-3de5c8388a96-kube-api-access-vml5s\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413595 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10e07288-52f4-407c-a8eb-709214e358f5-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413616 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8477e2b6-663f-49b8-a714-5925de307ac9-signing-cabundle\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413633 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/152126f1-6413-47f3-9847-ca6e3e561c7b-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413651 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-csi-data-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413667 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7923cd83-a73e-4a4b-9bdd-b38f7656579b-metrics-tls\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413693 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nnpvm\" (UniqueName: \"kubernetes.io/projected/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-kube-api-access-nnpvm\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413717 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/74bf2564-afaa-4967-9901-3fd670b5b991-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413739 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7923cd83-a73e-4a4b-9bdd-b38f7656579b-config-volume\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413757 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e599baf2-11e5-4d78-9282-b6850af2bd7e-service-ca-bundle\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413781 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwzw\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-kube-api-access-4nwzw\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.413799 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/023fe200-90c4-43c0-aa37-2241d9061e06-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.416501 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8477e2b6-663f-49b8-a714-5925de307ac9-signing-cabundle\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.416649 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c5ac55-d0d1-4075-84c6-3de5c8388a96-config\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.417823 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-certificates\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.423206 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97c5ac55-d0d1-4075-84c6-3de5c8388a96-serving-cert\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.405093 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-ca-trust-extracted\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.427524 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-tmp\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.428315 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.435571 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.435719 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-tls\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.436190 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7930ff4e-b5eb-42d7-ab81-884f18b423d7-config-volume\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.436684 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:30.936667315 +0000 UTC m=+114.969278837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.438611 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.439122 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.439699 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69e1d569-06c8-4098-8e3f-57c1a5970756-tmpfs\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.439722 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.440158 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8477e2b6-663f-49b8-a714-5925de307ac9-signing-key\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.440325 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.440407 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-dir\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.440481 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/152126f1-6413-47f3-9847-ca6e3e561c7b-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.441113 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fb509cc-d224-4492-8b34-b7625127411a-srv-cert\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.441248 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fb509cc-d224-4492-8b34-b7625127411a-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.441510 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb2106ad-091a-45e1-8e42-7dd01f174c03-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-jx99w\" (UID: \"eb2106ad-091a-45e1-8e42-7dd01f174c03\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.441773 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69e1d569-06c8-4098-8e3f-57c1a5970756-webhook-cert\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.442310 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.442625 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-installation-pull-secrets\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.443229 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/152126f1-6413-47f3-9847-ca6e3e561c7b-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.445218 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7930ff4e-b5eb-42d7-ab81-884f18b423d7-secret-volume\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.447321 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2337485-1832-41b3-97d8-4ccd31ed60a3-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-gg6pg\" (UID: \"e2337485-1832-41b3-97d8-4ccd31ed60a3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.447645 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-462wb"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.453972 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.454177 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.455206 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.456031 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.456198 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/33efc255-a7a4-4abf-93a8-a522763c8141-srv-cert\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.456571 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5tdt\" (UniqueName: \"kubernetes.io/projected/e2337485-1832-41b3-97d8-4ccd31ed60a3-kube-api-access-n5tdt\") pod \"package-server-manager-77f986bd66-gg6pg\" (UID: \"e2337485-1832-41b3-97d8-4ccd31ed60a3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.460351 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.462688 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/33efc255-a7a4-4abf-93a8-a522763c8141-profile-collector-cert\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.463236 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69e1d569-06c8-4098-8e3f-57c1a5970756-apiservice-cert\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.465874 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgqwt\" (UniqueName: \"kubernetes.io/projected/eb2106ad-091a-45e1-8e42-7dd01f174c03-kube-api-access-wgqwt\") pod \"cluster-samples-operator-6b564684c8-jx99w\" (UID: \"eb2106ad-091a-45e1-8e42-7dd01f174c03\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.467833 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.489762 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm7qv\" (UniqueName: \"kubernetes.io/projected/c2d80dcc-dd1a-434c-885d-16a7db229591-kube-api-access-vm7qv\") pod \"oauth-openshift-66458b6674-zpddx\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.515071 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516452 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5bx9q\" (UniqueName: \"kubernetes.io/projected/88ddb708-6eb5-4d13-ab7c-798b0a120818-kube-api-access-5bx9q\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516548 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-registration-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516584 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-mountpoint-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516611 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-stats-auth\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516634 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7mrk8\" (UniqueName: \"kubernetes.io/projected/e599baf2-11e5-4d78-9282-b6850af2bd7e-kube-api-access-7mrk8\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516664 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74bf2564-afaa-4967-9901-3fd670b5b991-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516686 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74bf2564-afaa-4967-9901-3fd670b5b991-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516728 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9w2pz\" (UniqueName: \"kubernetes.io/projected/f7edb507-fedd-4535-b7fe-46095dbd1b5c-kube-api-access-9w2pz\") pod \"ingress-canary-f8wb2\" (UID: \"f7edb507-fedd-4535-b7fe-46095dbd1b5c\") " pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.516756 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-socket-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517119 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n6kmt\" (UniqueName: \"kubernetes.io/projected/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-kube-api-access-n6kmt\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517192 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74bf2564-afaa-4967-9901-3fd670b5b991-config\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517230 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/023fe200-90c4-43c0-aa37-2241d9061e06-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517264 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/57ae9010-5764-463a-93ad-d4036ee2b53c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-xsmjk\" (UID: \"57ae9010-5764-463a-93ad-d4036ee2b53c\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517315 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rs7rs\" (UniqueName: \"kubernetes.io/projected/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-kube-api-access-rs7rs\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517337 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517359 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/10e07288-52f4-407c-a8eb-709214e358f5-ready\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517381 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-images\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517402 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10e07288-52f4-407c-a8eb-709214e358f5-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517422 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-default-certificate\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517457 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b68f5\" (UniqueName: \"kubernetes.io/projected/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-kube-api-access-b68f5\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517498 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517523 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-metrics-certs\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517546 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/023fe200-90c4-43c0-aa37-2241d9061e06-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517570 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-plugins-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517591 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4mb6c\" (UniqueName: \"kubernetes.io/projected/023fe200-90c4-43c0-aa37-2241d9061e06-kube-api-access-4mb6c\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517615 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fpds5\" (UniqueName: \"kubernetes.io/projected/7923cd83-a73e-4a4b-9bdd-b38f7656579b-kube-api-access-fpds5\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517635 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/88ddb708-6eb5-4d13-ab7c-798b0a120818-node-bootstrap-token\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517689 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517716 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f7edb507-fedd-4535-b7fe-46095dbd1b5c-cert\") pod \"ingress-canary-f8wb2\" (UID: \"f7edb507-fedd-4535-b7fe-46095dbd1b5c\") " pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517738 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7923cd83-a73e-4a4b-9bdd-b38f7656579b-tmp-dir\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517764 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517803 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2jm87\" (UniqueName: \"kubernetes.io/projected/10e07288-52f4-407c-a8eb-709214e358f5-kube-api-access-2jm87\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.517827 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5kpbh\" (UniqueName: \"kubernetes.io/projected/9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd-kube-api-access-5kpbh\") pod \"migrator-866fcbc849-xkvs8\" (UID: \"9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518008 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ht2w7\" (UniqueName: \"kubernetes.io/projected/57ae9010-5764-463a-93ad-d4036ee2b53c-kube-api-access-ht2w7\") pod \"control-plane-machine-set-operator-75ffdb6fcd-xsmjk\" (UID: \"57ae9010-5764-463a-93ad-d4036ee2b53c\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518086 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10e07288-52f4-407c-a8eb-709214e358f5-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518113 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-csi-data-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518128 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7923cd83-a73e-4a4b-9bdd-b38f7656579b-metrics-tls\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518150 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/74bf2564-afaa-4967-9901-3fd670b5b991-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518171 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7923cd83-a73e-4a4b-9bdd-b38f7656579b-config-volume\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518192 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e599baf2-11e5-4d78-9282-b6850af2bd7e-service-ca-bundle\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518216 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/023fe200-90c4-43c0-aa37-2241d9061e06-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.518234 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/88ddb708-6eb5-4d13-ab7c-798b0a120818-certs\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.521371 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.02134848 +0000 UTC m=+115.053959992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.521697 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-registration-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.521743 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-mountpoint-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.523167 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/023fe200-90c4-43c0-aa37-2241d9061e06-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.524995 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/88ddb708-6eb5-4d13-ab7c-798b0a120818-certs\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.529000 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-socket-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.529463 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/10e07288-52f4-407c-a8eb-709214e358f5-ready\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.529736 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74bf2564-afaa-4967-9901-3fd670b5b991-config\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.530195 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-csi-data-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.530907 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10e07288-52f4-407c-a8eb-709214e358f5-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.531312 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-plugins-dir\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.531969 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.533315 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4tsw\" (UniqueName: \"kubernetes.io/projected/5fb509cc-d224-4492-8b34-b7625127411a-kube-api-access-z4tsw\") pod \"catalog-operator-75ff9f647d-gkd64\" (UID: \"5fb509cc-d224-4492-8b34-b7625127411a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.536105 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7923cd83-a73e-4a4b-9bdd-b38f7656579b-config-volume\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.536120 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.536465 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/74bf2564-afaa-4967-9901-3fd670b5b991-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.536664 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10e07288-52f4-407c-a8eb-709214e358f5-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.539996 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7923cd83-a73e-4a4b-9bdd-b38f7656579b-tmp-dir\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.542944 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-images\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.544294 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e599baf2-11e5-4d78-9282-b6850af2bd7e-service-ca-bundle\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.544417 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/88ddb708-6eb5-4d13-ab7c-798b0a120818-node-bootstrap-token\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.545651 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-default-certificate\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.545790 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f7edb507-fedd-4535-b7fe-46095dbd1b5c-cert\") pod \"ingress-canary-f8wb2\" (UID: \"f7edb507-fedd-4535-b7fe-46095dbd1b5c\") " pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.545972 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7923cd83-a73e-4a4b-9bdd-b38f7656579b-metrics-tls\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.552072 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnpvm\" (UniqueName: \"kubernetes.io/projected/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-kube-api-access-nnpvm\") pod \"marketplace-operator-547dbd544d-95rbr\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.552673 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vml5s\" (UniqueName: \"kubernetes.io/projected/97c5ac55-d0d1-4075-84c6-3de5c8388a96-kube-api-access-vml5s\") pod \"service-ca-operator-5b9c976747-lhdwb\" (UID: \"97c5ac55-d0d1-4075-84c6-3de5c8388a96\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.552782 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.553747 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/57ae9010-5764-463a-93ad-d4036ee2b53c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-xsmjk\" (UID: \"57ae9010-5764-463a-93ad-d4036ee2b53c\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.555149 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-stats-auth\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.555348 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/023fe200-90c4-43c0-aa37-2241d9061e06-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.555656 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e599baf2-11e5-4d78-9282-b6850af2bd7e-metrics-certs\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.557379 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: W0317 18:49:30.560893 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1c85da3_d4ca_4abe_b757_1df684b37eb8.slice/crio-a582108c8f62e16992f9b7fb393caca8c182a1f050a3194b8cfec1e771f1ddd2 WatchSource:0}: Error finding container a582108c8f62e16992f9b7fb393caca8c182a1f050a3194b8cfec1e771f1ddd2: Status 404 returned error can't find the container with id a582108c8f62e16992f9b7fb393caca8c182a1f050a3194b8cfec1e771f1ddd2 Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.608574 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74bf2564-afaa-4967-9901-3fd670b5b991-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.613184 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8w7l\" (UniqueName: \"kubernetes.io/projected/33efc255-a7a4-4abf-93a8-a522763c8141-kube-api-access-k8w7l\") pod \"olm-operator-5cdf44d969-h46kj\" (UID: \"33efc255-a7a4-4abf-93a8-a522763c8141\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.620722 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-bound-sa-token\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.622299 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.622672 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.122658778 +0000 UTC m=+115.155270300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.633276 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-67c89758df-tnkj2"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.639871 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwzw\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-kube-api-access-4nwzw\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.646667 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-h787x"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.649946 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gvc\" (UniqueName: \"kubernetes.io/projected/8477e2b6-663f-49b8-a714-5925de307ac9-kube-api-access-57gvc\") pod \"service-ca-74545575db-lz8zr\" (UID: \"8477e2b6-663f-49b8-a714-5925de307ac9\") " pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.665428 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cmlt\" (UniqueName: \"kubernetes.io/projected/69e1d569-06c8-4098-8e3f-57c1a5970756-kube-api-access-9cmlt\") pod \"packageserver-7d4fc7d867-k7cdw\" (UID: \"69e1d569-06c8-4098-8e3f-57c1a5970756\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.690297 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.690856 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/152126f1-6413-47f3-9847-ca6e3e561c7b-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-s79f7\" (UID: \"152126f1-6413-47f3-9847-ca6e3e561c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.705429 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-srk6n\" (UniqueName: \"kubernetes.io/projected/7930ff4e-b5eb-42d7-ab81-884f18b423d7-kube-api-access-srk6n\") pod \"collect-profiles-29562885-nwdmv\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.711113 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.717303 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.723214 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.723725 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.223705659 +0000 UTC m=+115.256317181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.723740 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.726790 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b68f5\" (UniqueName: \"kubernetes.io/projected/e7e11c8a-b0f2-4f43-bc25-43cb7f57f952-kube-api-access-b68f5\") pod \"machine-config-controller-f9cdd68f7-8gfln\" (UID: \"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.729703 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.742475 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-74545575db-lz8zr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.748805 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bx9q\" (UniqueName: \"kubernetes.io/projected/88ddb708-6eb5-4d13-ab7c-798b0a120818-kube-api-access-5bx9q\") pod \"machine-config-server-455vz\" (UID: \"88ddb708-6eb5-4d13-ab7c-798b0a120818\") " pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.749495 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.756966 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.767835 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.771839 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mrk8\" (UniqueName: \"kubernetes.io/projected/e599baf2-11e5-4d78-9282-b6850af2bd7e-kube-api-access-7mrk8\") pod \"router-default-68cf44c8b8-vhklb\" (UID: \"e599baf2-11e5-4d78-9282-b6850af2bd7e\") " pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.773810 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.800619 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.810374 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6kmt\" (UniqueName: \"kubernetes.io/projected/54a6c559-52e9-4f87-88f6-0d0ab99aeebc-kube-api-access-n6kmt\") pod \"machine-config-operator-67c9d58cbb-5g6pb\" (UID: \"54a6c559-52e9-4f87-88f6-0d0ab99aeebc\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.812345 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.819728 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74bf2564-afaa-4967-9901-3fd670b5b991-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-bn9fl\" (UID: \"74bf2564-afaa-4967-9901-3fd670b5b991\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.820083 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.827079 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.327046414 +0000 UTC m=+115.359657936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.834485 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w2pz\" (UniqueName: \"kubernetes.io/projected/f7edb507-fedd-4535-b7fe-46095dbd1b5c-kube-api-access-9w2pz\") pod \"ingress-canary-f8wb2\" (UID: \"f7edb507-fedd-4535-b7fe-46095dbd1b5c\") " pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.835522 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.846035 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs7rs\" (UniqueName: \"kubernetes.io/projected/779e1c5c-8a98-4f8d-91f5-3cdf72ac1888-kube-api-access-rs7rs\") pod \"csi-hostpathplugin-ddn25\" (UID: \"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888\") " pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.846806 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-455vz" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.865921 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jm87\" (UniqueName: \"kubernetes.io/projected/10e07288-52f4-407c-a8eb-709214e358f5-kube-api-access-2jm87\") pod \"cni-sysctl-allowlist-ds-jqrnn\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.873355 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn"] Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.886453 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.896903 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-f8wb2" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.903953 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kpbh\" (UniqueName: \"kubernetes.io/projected/9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd-kube-api-access-5kpbh\") pod \"migrator-866fcbc849-xkvs8\" (UID: \"9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.913901 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht2w7\" (UniqueName: \"kubernetes.io/projected/57ae9010-5764-463a-93ad-d4036ee2b53c-kube-api-access-ht2w7\") pod \"control-plane-machine-set-operator-75ffdb6fcd-xsmjk\" (UID: \"57ae9010-5764-463a-93ad-d4036ee2b53c\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.936746 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:30 crc kubenswrapper[5110]: E0317 18:49:30.937154 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.437136136 +0000 UTC m=+115.469747658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:30 crc kubenswrapper[5110]: I0317 18:49:30.975616 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpds5\" (UniqueName: \"kubernetes.io/projected/7923cd83-a73e-4a4b-9bdd-b38f7656579b-kube-api-access-fpds5\") pod \"dns-default-dhxm4\" (UID: \"7923cd83-a73e-4a4b-9bdd-b38f7656579b\") " pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.003728 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/023fe200-90c4-43c0-aa37-2241d9061e06-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.003820 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.004330 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mb6c\" (UniqueName: \"kubernetes.io/projected/023fe200-90c4-43c0-aa37-2241d9061e06-kube-api-access-4mb6c\") pod \"ingress-operator-6b9cb4dbcf-gzdmn\" (UID: \"023fe200-90c4-43c0-aa37-2241d9061e06\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.035956 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-z7gg7"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.038139 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.038835 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.538821545 +0000 UTC m=+115.571433067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.093387 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.099448 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.100194 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.104266 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.122262 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-747b44746d-86ddn"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.130745 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.137238 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.139590 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.139715 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.639696221 +0000 UTC m=+115.672307743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.140018 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.140556 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.640544675 +0000 UTC m=+115.673156197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.143862 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-wcbht"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.180743 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.185255 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-zpddx"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.185492 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" event={"ID":"84f8e2e3-408f-4873-90e9-f778b20e1a32","Type":"ContainerStarted","Data":"4da29f11bfe271ef7f8ff69b2b8446504ad95d1bc38d8e8ec646db0f4b62807e"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.185999 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d44f6ddf-8jwk2"] Mar 17 18:49:31 crc kubenswrapper[5110]: W0317 18:49:31.188973 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0a5117c_465a_4abf_830d_fac77933f27c.slice/crio-6660d96fe9d9df533433ef4d3400544b12cae22e08f38cfa517d688fdef8c9cd WatchSource:0}: Error finding container 6660d96fe9d9df533433ef4d3400544b12cae22e08f38cfa517d688fdef8c9cd: Status 404 returned error can't find the container with id 6660d96fe9d9df533433ef4d3400544b12cae22e08f38cfa517d688fdef8c9cd Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.194966 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.200646 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.203536 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" event={"ID":"6b5d0e6a-c5e4-409b-a868-61480630c906","Type":"ContainerStarted","Data":"21185fd4bd8ec1fde191b47b0b53d9257ffbe20ab63cfef88037ed8f2de841b0"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.203615 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" event={"ID":"6b5d0e6a-c5e4-409b-a868-61480630c906","Type":"ContainerStarted","Data":"76a2051925090139d6e3a36df7c4ae27878cc1a40fc9ae700736c9a401b1c293"} Mar 17 18:49:31 crc kubenswrapper[5110]: W0317 18:49:31.214199 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23f9945f_bddd_4c0a_bf2c_cc0baee81974.slice/crio-98f76482ec678d73d0c0476480d9482cf3762d70b37a9d6f708fc616eaa13ffa WatchSource:0}: Error finding container 98f76482ec678d73d0c0476480d9482cf3762d70b37a9d6f708fc616eaa13ffa: Status 404 returned error can't find the container with id 98f76482ec678d73d0c0476480d9482cf3762d70b37a9d6f708fc616eaa13ffa Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.217817 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" event={"ID":"e1c85da3-d4ca-4abe-b757-1df684b37eb8","Type":"ContainerStarted","Data":"6745e5d77074f4ce2df820503cc74fe7cfd9d5af6c66b649fcb287fe18b1ebec"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.217873 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" event={"ID":"e1c85da3-d4ca-4abe-b757-1df684b37eb8","Type":"ContainerStarted","Data":"a582108c8f62e16992f9b7fb393caca8c182a1f050a3194b8cfec1e771f1ddd2"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.225313 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" event={"ID":"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196","Type":"ContainerStarted","Data":"6edaf95a2cb1e9ee75804127c9225171aa1f783c530b36882f142e6c5db629b0"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.225351 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" event={"ID":"c8f3b28c-0ef5-4d62-aff9-4d8ad1a9d196","Type":"ContainerStarted","Data":"3c2c9f9afa7be8fc5c3f1ecc0131aa470ae84406d5826c50efdfc801c2131302"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.230604 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" event={"ID":"1bda76e2-be33-4ca9-b962-c64fa6d4df1e","Type":"ContainerStarted","Data":"39c318266571a4862abb25dadd39d7faa1af36b0f37a934380c33b80a88fd47c"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.230644 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" event={"ID":"1bda76e2-be33-4ca9-b962-c64fa6d4df1e","Type":"ContainerStarted","Data":"850382f91ce7ec354e94d4f0a86e96bc5dd9ac8a3c366b02ace76f3210ccc146"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.243656 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.244495 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.744450905 +0000 UTC m=+115.777062427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: W0317 18:49:31.261522 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab45c12d_7f9a_4053_b810_645c966906a1.slice/crio-f342c37e5096aca1f6d8913c81db013d1855a877ae3b4bcab7b09d67adcd54d0 WatchSource:0}: Error finding container f342c37e5096aca1f6d8913c81db013d1855a877ae3b4bcab7b09d67adcd54d0: Status 404 returned error can't find the container with id f342c37e5096aca1f6d8913c81db013d1855a877ae3b4bcab7b09d67adcd54d0 Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.267749 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" event={"ID":"5b9a2661-d727-4c6b-97cc-7d52281d9425","Type":"ContainerStarted","Data":"85d491b08d0717497453e803478b85afd77de435b33380e3fa1e398cc5a4e2b2"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.287572 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" event={"ID":"5e9bdc21-2fd6-43b9-9666-50637fef0592","Type":"ContainerStarted","Data":"61e739545b34035961b30ba649f434f940d19275628e8e4984675c896173db7e"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.304038 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" event={"ID":"54c98382-43ad-4113-ab43-b31eb8d4bc58","Type":"ContainerStarted","Data":"e94d12d728725140b6cfdde22f2e60703e07ced23af326ceae517eb8109dbcee"} Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.322660 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.351873 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.357615 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.857598503 +0000 UTC m=+115.890210025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.392866 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" podStartSLOduration=93.392850963 podStartE2EDuration="1m33.392850963s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:31.390282782 +0000 UTC m=+115.422894304" watchObservedRunningTime="2026-03-17 18:49:31.392850963 +0000 UTC m=+115.425462485" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.425914 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.444233 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-74545575db-lz8zr"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.465969 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.466310 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:31.966293636 +0000 UTC m=+115.998905158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: W0317 18:49:31.466829 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33efc255_a7a4_4abf_93a8_a522763c8141.slice/crio-cc4ef78e7e2c457aab087009ce8da854519fb48a224d71e163c2ca65b9cc9728 WatchSource:0}: Error finding container cc4ef78e7e2c457aab087009ce8da854519fb48a224d71e163c2ca65b9cc9728: Status 404 returned error can't find the container with id cc4ef78e7e2c457aab087009ce8da854519fb48a224d71e163c2ca65b9cc9728 Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.550975 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-bnwct" podStartSLOduration=94.550957011 podStartE2EDuration="1m34.550957011s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:31.549041628 +0000 UTC m=+115.581653150" watchObservedRunningTime="2026-03-17 18:49:31.550957011 +0000 UTC m=+115.583568533" Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.574384 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.574742 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.074726092 +0000 UTC m=+116.107337624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: W0317 18:49:31.578784 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8477e2b6_663f_49b8_a714_5925de307ac9.slice/crio-4957e1d22aac95dee208c773d7f973f5bfee4dc8ec5ef7e7047121eea0bc5e9b WatchSource:0}: Error finding container 4957e1d22aac95dee208c773d7f973f5bfee4dc8ec5ef7e7047121eea0bc5e9b: Status 404 returned error can't find the container with id 4957e1d22aac95dee208c773d7f973f5bfee4dc8ec5ef7e7047121eea0bc5e9b Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.629818 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.671659 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-95rbr"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.674128 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.676291 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.676674 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.176654537 +0000 UTC m=+116.209266059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.682110 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.683321 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.709943 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.716980 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-f8wb2"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.724232 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.780781 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.781104 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.281091352 +0000 UTC m=+116.313702874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.826253 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.856862 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.857192 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.883803 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.884255 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.38423719 +0000 UTC m=+116.416848712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.930953 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ddn25"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.950641 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dhxm4"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.955585 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8"] Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.985174 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:31 crc kubenswrapper[5110]: E0317 18:49:31.985780 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.485767255 +0000 UTC m=+116.518378777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:31 crc kubenswrapper[5110]: I0317 18:49:31.992283 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk"] Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.020318 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-ltfln" podStartSLOduration=94.020303485 podStartE2EDuration="1m34.020303485s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.018985079 +0000 UTC m=+116.051596601" watchObservedRunningTime="2026-03-17 18:49:32.020303485 +0000 UTC m=+116.052915007" Mar 17 18:49:32 crc kubenswrapper[5110]: W0317 18:49:32.067082 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29836d6d_6ed4_4ed6_9de6_303f7bf98ab5.slice/crio-68f777267fb4ec1644dd1ad144840b6905cfe1a4c3e41ce8c8d77b29e0b4bb97 WatchSource:0}: Error finding container 68f777267fb4ec1644dd1ad144840b6905cfe1a4c3e41ce8c8d77b29e0b4bb97: Status 404 returned error can't find the container with id 68f777267fb4ec1644dd1ad144840b6905cfe1a4c3e41ce8c8d77b29e0b4bb97 Mar 17 18:49:32 crc kubenswrapper[5110]: W0317 18:49:32.073977 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e11c8a_b0f2_4f43_bc25_43cb7f57f952.slice/crio-c67557d46092c9cd5774dfdf5027318ad7a721ea7c905932a9022733208b4dd9 WatchSource:0}: Error finding container c67557d46092c9cd5774dfdf5027318ad7a721ea7c905932a9022733208b4dd9: Status 404 returned error can't find the container with id c67557d46092c9cd5774dfdf5027318ad7a721ea7c905932a9022733208b4dd9 Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.089826 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.090166 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.590148188 +0000 UTC m=+116.622759710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: W0317 18:49:32.101299 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7edb507_fedd_4535_b7fe_46095dbd1b5c.slice/crio-49211deac351cc4e024edd7ce393d44a5a2a62d04d5aede1854f9b5a55754ac1 WatchSource:0}: Error finding container 49211deac351cc4e024edd7ce393d44a5a2a62d04d5aede1854f9b5a55754ac1: Status 404 returned error can't find the container with id 49211deac351cc4e024edd7ce393d44a5a2a62d04d5aede1854f9b5a55754ac1 Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.109205 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl"] Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.116317 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" podStartSLOduration=94.116299286 podStartE2EDuration="1m34.116299286s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.115045861 +0000 UTC m=+116.147657383" watchObservedRunningTime="2026-03-17 18:49:32.116299286 +0000 UTC m=+116.148910828" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.191525 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.191870 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.691846967 +0000 UTC m=+116.724458489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: W0317 18:49:32.237979 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b20488e_b2ad_4a1d_83e7_18c7e27ca8bd.slice/crio-ea8a9644e047ab9431971ecd711969e8af7f7874a59bfd1631403e0b83768430 WatchSource:0}: Error finding container ea8a9644e047ab9431971ecd711969e8af7f7874a59bfd1631403e0b83768430: Status 404 returned error can't find the container with id ea8a9644e047ab9431971ecd711969e8af7f7874a59bfd1631403e0b83768430 Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.262032 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-vnjj4" podStartSLOduration=95.262011079 podStartE2EDuration="1m35.262011079s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.227756556 +0000 UTC m=+116.260368128" watchObservedRunningTime="2026-03-17 18:49:32.262011079 +0000 UTC m=+116.294622601" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.293649 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.293802 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.793765752 +0000 UTC m=+116.826377274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.294003 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.294441 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.794431411 +0000 UTC m=+116.827042943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.345233 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" podStartSLOduration=95.345205343 podStartE2EDuration="1m35.345205343s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.299754179 +0000 UTC m=+116.332365711" watchObservedRunningTime="2026-03-17 18:49:32.345205343 +0000 UTC m=+116.377816875" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.385241 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" event={"ID":"7930ff4e-b5eb-42d7-ab81-884f18b423d7","Type":"ContainerStarted","Data":"cc8bf6b8dda570da86086544d743910903afd53a89a3a04cc08313eac3def8dd"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.395397 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" event={"ID":"152126f1-6413-47f3-9847-ca6e3e561c7b","Type":"ContainerStarted","Data":"43b2f9bf22a99ceefacc08bb73de8187a61865bd22272e6d25556cffff69cd6c"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.395917 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.396279 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.896259333 +0000 UTC m=+116.928870865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.399184 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" event={"ID":"ab45c12d-7f9a-4053-b810-645c966906a1","Type":"ContainerStarted","Data":"f342c37e5096aca1f6d8913c81db013d1855a877ae3b4bcab7b09d67adcd54d0"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.432673 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" event={"ID":"10e07288-52f4-407c-a8eb-709214e358f5","Type":"ContainerStarted","Data":"c33605948ab99822eeeef3dad82ac7fe459817fb201ec525b04509b9dd961700"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.437719 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-455vz" event={"ID":"88ddb708-6eb5-4d13-ab7c-798b0a120818","Type":"ContainerStarted","Data":"e4733a220c95b099fdb530ed66106da9d09fd8e115d4c7abea224ac71af6e451"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.444183 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" event={"ID":"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5","Type":"ContainerStarted","Data":"68f777267fb4ec1644dd1ad144840b6905cfe1a4c3e41ce8c8d77b29e0b4bb97"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.478466 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" event={"ID":"33efc255-a7a4-4abf-93a8-a522763c8141","Type":"ContainerStarted","Data":"cc4ef78e7e2c457aab087009ce8da854519fb48a224d71e163c2ca65b9cc9728"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.480513 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" event={"ID":"c0a5117c-465a-4abf-830d-fac77933f27c","Type":"ContainerStarted","Data":"f990166c17ad7ffc6572f5735c3e37a8d2aeab03769fc3d420b73b7a2e8b197f"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.480545 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" event={"ID":"c0a5117c-465a-4abf-830d-fac77933f27c","Type":"ContainerStarted","Data":"6660d96fe9d9df533433ef4d3400544b12cae22e08f38cfa517d688fdef8c9cd"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.494639 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-747b44746d-86ddn" event={"ID":"9731f229-c1b8-4511-819c-b90e6cd6af2e","Type":"ContainerStarted","Data":"c0a05e12ff1c24499cb366843d9ea74a8a4c44dcb9059fe9938b7789a19457b2"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.494683 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-747b44746d-86ddn" event={"ID":"9731f229-c1b8-4511-819c-b90e6cd6af2e","Type":"ContainerStarted","Data":"1bb89f61e8e871c5a1082ebc6c175d0ed906efd18545fbbf2d086f208428325b"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.494981 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.496136 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" event={"ID":"5e9bdc21-2fd6-43b9-9666-50637fef0592","Type":"ContainerStarted","Data":"bbe0159d631faf7e4e3ac7bf0dfc5c369c89e4fe13d3304ecfe75edf780868a8"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.496860 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.497366 5110 patch_prober.go:28] interesting pod/downloads-747b44746d-86ddn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.497404 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-86ddn" podUID="9731f229-c1b8-4511-819c-b90e6cd6af2e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.497458 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.497733 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:32.997721605 +0000 UTC m=+117.030333127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.503626 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" event={"ID":"023fe200-90c4-43c0-aa37-2241d9061e06","Type":"ContainerStarted","Data":"9497ba50b66c66307415fc15ec867326e8d0edaaafa65b9f3b91a7511d59cb7c"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.510972 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d44f6ddf-8jwk2" event={"ID":"23f9945f-bddd-4c0a-bf2c-cc0baee81974","Type":"ContainerStarted","Data":"98f76482ec678d73d0c0476480d9482cf3762d70b37a9d6f708fc616eaa13ffa"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.511872 5110 patch_prober.go:28] interesting pod/console-operator-67c89758df-tnkj2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.511919 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" podUID="5e9bdc21-2fd6-43b9-9666-50637fef0592" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.524329 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" event={"ID":"54c98382-43ad-4113-ab43-b31eb8d4bc58","Type":"ContainerStarted","Data":"7f7f07cad683a8c5f9b31786c7188b5bdc90e761ad3453c7ffd9cee99b144b61"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.527179 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" event={"ID":"e599baf2-11e5-4d78-9282-b6850af2bd7e","Type":"ContainerStarted","Data":"84fb927bada63d0de9b21b5b8e6c95091dc0b17697fcf5f322f18831528c98b2"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.533186 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" event={"ID":"c2d80dcc-dd1a-434c-885d-16a7db229591","Type":"ContainerStarted","Data":"aea08505f36332454deb8437feb446a7b205125aaac2f39f4cda929bcf73116c"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.533221 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" event={"ID":"c2d80dcc-dd1a-434c-885d-16a7db229591","Type":"ContainerStarted","Data":"9709f9ca1616996feb5d65ff651a92b74e19640ae59c1c31ccb434e0d38c0412"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.534153 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.535133 5110 patch_prober.go:28] interesting pod/oauth-openshift-66458b6674-zpddx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" start-of-body= Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.535172 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.536885 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" event={"ID":"1bda76e2-be33-4ca9-b962-c64fa6d4df1e","Type":"ContainerStarted","Data":"ec6b8c6363d1ac3b866052dcef2594e3f7753e0a35d124c78486640b31299e87"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.539275 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" event={"ID":"57ae9010-5764-463a-93ad-d4036ee2b53c","Type":"ContainerStarted","Data":"4e7a16a5d3bce7295a621d81b15cb17b23c5a6b73d4a25776269d5a60ce45993"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.540121 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-f8wb2" event={"ID":"f7edb507-fedd-4535-b7fe-46095dbd1b5c","Type":"ContainerStarted","Data":"49211deac351cc4e024edd7ce393d44a5a2a62d04d5aede1854f9b5a55754ac1"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.543368 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" event={"ID":"5fb509cc-d224-4492-8b34-b7625127411a","Type":"ContainerStarted","Data":"f83d345cf7431806f45dc16f116de7f9b03b506e54440248b314f5b56d74e8d2"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.546446 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-755bb95488-xf26c" podStartSLOduration=94.54642734 podStartE2EDuration="1m34.54642734s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.501005787 +0000 UTC m=+116.533617309" watchObservedRunningTime="2026-03-17 18:49:32.54642734 +0000 UTC m=+116.579038862" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.557839 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" event={"ID":"967f9e08-ced9-444e-b03a-439f0f4cc686","Type":"ContainerStarted","Data":"33bf096ab4ddf9d08e0e08fb0083506f1b2f504ab63bcec7cf04497ac3f4c260"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.562273 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" event={"ID":"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888","Type":"ContainerStarted","Data":"3242800bf0d52cf267c5c7c7e7b4975e702b0a65711b80bf331190cb0edb612a"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.564464 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" event={"ID":"e2337485-1832-41b3-97d8-4ccd31ed60a3","Type":"ContainerStarted","Data":"bd9a682e99d186e0f041cd17c5b9056722fa0f7f36c02393e4b1900e6e0d2e4e"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.564514 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" event={"ID":"e2337485-1832-41b3-97d8-4ccd31ed60a3","Type":"ContainerStarted","Data":"a4f2e58f18064183753087b6cf59653c47a3e264dae988b8eda32a93466ca6cf"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.565864 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" event={"ID":"9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd","Type":"ContainerStarted","Data":"ea8a9644e047ab9431971ecd711969e8af7f7874a59bfd1631403e0b83768430"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.566820 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" event={"ID":"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952","Type":"ContainerStarted","Data":"c67557d46092c9cd5774dfdf5027318ad7a721ea7c905932a9022733208b4dd9"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.571168 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" event={"ID":"deb743a6-6308-466e-830a-ae0a66472896","Type":"ContainerStarted","Data":"b29b9cfb57acbb3e5bd9303a503e93d2e2b247b50da7287c7b0e4e75939acaae"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.579936 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" event={"ID":"6b5d0e6a-c5e4-409b-a868-61480630c906","Type":"ContainerStarted","Data":"d9603c7b0fadc8dac41a7a22a0bca4c47c5aa6cbf13966de77c5f717cde1c421"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.591682 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" event={"ID":"97c5ac55-d0d1-4075-84c6-3de5c8388a96","Type":"ContainerStarted","Data":"c01bbe42885b4484f84073195b90e5654be1c00753d8ced1b365a65fef95107e"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.598740 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.602972 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" event={"ID":"74bf2564-afaa-4967-9901-3fd670b5b991","Type":"ContainerStarted","Data":"370fe1b212c4972e67b3cfc8ccf10ea10b2a4155c23dee6d9a9cd602fe18c060"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.605725 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-74545575db-lz8zr" event={"ID":"8477e2b6-663f-49b8-a714-5925de307ac9","Type":"ContainerStarted","Data":"4957e1d22aac95dee208c773d7f973f5bfee4dc8ec5ef7e7047121eea0bc5e9b"} Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.629701 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.129672296 +0000 UTC m=+117.162283828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.647996 5110 generic.go:358] "Generic (PLEG): container finished" podID="84f8e2e3-408f-4873-90e9-f778b20e1a32" containerID="19951d6fc02b25d16ada03cd8d1e3ab24586bedba2084fc9bb1c5700d7af120f" exitCode=0 Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.648362 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" event={"ID":"84f8e2e3-408f-4873-90e9-f778b20e1a32","Type":"ContainerDied","Data":"19951d6fc02b25d16ada03cd8d1e3ab24586bedba2084fc9bb1c5700d7af120f"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.670541 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" event={"ID":"54a6c559-52e9-4f87-88f6-0d0ab99aeebc","Type":"ContainerStarted","Data":"a1a5570c1f1f090afb01c33d55b835a7af712e078818bfb55d037a2ba63c6fb2"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.686870 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" event={"ID":"eb2106ad-091a-45e1-8e42-7dd01f174c03","Type":"ContainerStarted","Data":"099ddec70bce25fd6a79eb9f5332b3b2e905839b10c732f2d6b9d3af165dfa7a"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.721134 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" event={"ID":"f8606fbd-7611-4970-9f65-f3e12bb76894","Type":"ContainerStarted","Data":"851bd270b5035abfe4125bd73c3a1282b29135bbc64037b92e0caa6cebb872fc"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.726076 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dhxm4" event={"ID":"7923cd83-a73e-4a4b-9bdd-b38f7656579b","Type":"ContainerStarted","Data":"8d4329df3f66058d5f8e6f3d65dea13e672d98080c81c9ec61784a4ea9eb4e26"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.728101 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" event={"ID":"69e1d569-06c8-4098-8e3f-57c1a5970756","Type":"ContainerStarted","Data":"dbff35cef8d33d518471e25bb5f52bf0e6ab731f96771066a8121f7cbb2544bf"} Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.732021 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.734973 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.234956224 +0000 UTC m=+117.267567756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.762969 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-799b87ffcd-875ck" podStartSLOduration=95.762950813 podStartE2EDuration="1m35.762950813s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.760714931 +0000 UTC m=+116.793326453" watchObservedRunningTime="2026-03-17 18:49:32.762950813 +0000 UTC m=+116.795562335" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.849618 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.851109 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.351091835 +0000 UTC m=+117.383703357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.876273 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-54c688565-hhs4z" podStartSLOduration=96.876252765 podStartE2EDuration="1m36.876252765s" podCreationTimestamp="2026-03-17 18:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.875863584 +0000 UTC m=+116.908475106" watchObservedRunningTime="2026-03-17 18:49:32.876252765 +0000 UTC m=+116.908864287" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.888667 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" podStartSLOduration=95.888644189 podStartE2EDuration="1m35.888644189s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.786755075 +0000 UTC m=+116.819366597" watchObservedRunningTime="2026-03-17 18:49:32.888644189 +0000 UTC m=+116.921255711" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.925503 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" podStartSLOduration=95.925486244 podStartE2EDuration="1m35.925486244s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.925169145 +0000 UTC m=+116.957780677" watchObservedRunningTime="2026-03-17 18:49:32.925486244 +0000 UTC m=+116.958097766" Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.954473 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:32 crc kubenswrapper[5110]: E0317 18:49:32.955168 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.455151209 +0000 UTC m=+117.487762731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:32 crc kubenswrapper[5110]: I0317 18:49:32.979464 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-69b85846b6-462wb" podStartSLOduration=95.979440085 podStartE2EDuration="1m35.979440085s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.941711025 +0000 UTC m=+116.974322547" watchObservedRunningTime="2026-03-17 18:49:32.979440085 +0000 UTC m=+117.012051607" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.027887 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-f8xnn" podStartSLOduration=96.027695097 podStartE2EDuration="1m36.027695097s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:32.981027039 +0000 UTC m=+117.013638571" watchObservedRunningTime="2026-03-17 18:49:33.027695097 +0000 UTC m=+117.060306619" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.029364 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-747b44746d-86ddn" podStartSLOduration=96.029352953 podStartE2EDuration="1m36.029352953s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:33.027721908 +0000 UTC m=+117.060333450" watchObservedRunningTime="2026-03-17 18:49:33.029352953 +0000 UTC m=+117.061964475" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.057365 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.057680 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.557663111 +0000 UTC m=+117.590274633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.191091 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.191402 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.69138985 +0000 UTC m=+117.724001372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.292002 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.292400 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.792347079 +0000 UTC m=+117.824958591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.293508 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.294340 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.794319324 +0000 UTC m=+117.826930846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.426156 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.426642 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:33.926621294 +0000 UTC m=+117.959232816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.529343 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.529719 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.029707881 +0000 UTC m=+118.062319403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.591962 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.592014 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.605288 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.645031 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.645172 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.145143982 +0000 UTC m=+118.177755504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.645450 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.645888 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.145868012 +0000 UTC m=+118.178479534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.744264 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-455vz" event={"ID":"88ddb708-6eb5-4d13-ab7c-798b0a120818","Type":"ContainerStarted","Data":"a8b5fcaf6f92ff2a2de65f8a0997db49ae066df6d8cfd9a2ec1d70be0fd91b28"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.750666 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.751138 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.25111809 +0000 UTC m=+118.283729612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.752188 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" event={"ID":"33efc255-a7a4-4abf-93a8-a522763c8141","Type":"ContainerStarted","Data":"bef899a5d110253491a108990317589f3ba8315fab3b123f7fed17866d1523c1"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.755651 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" event={"ID":"023fe200-90c4-43c0-aa37-2241d9061e06","Type":"ContainerStarted","Data":"d27a0958a5edc65b8eea0b895b82f93874c758fced59f0d1673e6ccd1cfd2d68"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.764514 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d44f6ddf-8jwk2" event={"ID":"23f9945f-bddd-4c0a-bf2c-cc0baee81974","Type":"ContainerStarted","Data":"2e775307b0ed5b54f6728c830003d7d74885a634e3cfca7b476accad685b63ac"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.766890 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" event={"ID":"e599baf2-11e5-4d78-9282-b6850af2bd7e","Type":"ContainerStarted","Data":"bb28c344ebddb3691fa8c0925c0c4ccea0888579e2eecccc142f82aacc63c4ab"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.770801 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.772242 5110 patch_prober.go:28] interesting pod/olm-operator-5cdf44d969-h46kj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.772287 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" podUID="33efc255-a7a4-4abf-93a8-a522763c8141" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.801491 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.806200 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:33 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:33 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:33 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.806250 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.826989 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-f8wb2" event={"ID":"f7edb507-fedd-4535-b7fe-46095dbd1b5c","Type":"ContainerStarted","Data":"d53764102b7f830d358163f6f4703833e09c98e1d844f21f3aae56459fc3886f"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.843586 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-455vz" podStartSLOduration=6.843555971 podStartE2EDuration="6.843555971s" podCreationTimestamp="2026-03-17 18:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:33.837881513 +0000 UTC m=+117.870493035" watchObservedRunningTime="2026-03-17 18:49:33.843555971 +0000 UTC m=+117.876167493" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.850014 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" event={"ID":"5fb509cc-d224-4492-8b34-b7625127411a","Type":"ContainerStarted","Data":"9045dcf8e7cda0043b56b383d537a739c35b2d78a26ee6698e86806962a3a19e"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.850811 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.854524 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:33 crc kubenswrapper[5110]: E0317 18:49:33.858011 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.357995563 +0000 UTC m=+118.390607075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.859030 5110 patch_prober.go:28] interesting pod/catalog-operator-75ff9f647d-gkd64 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.859196 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" podUID="5fb509cc-d224-4492-8b34-b7625127411a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.861583 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64d44f6ddf-8jwk2" podStartSLOduration=96.861563822 podStartE2EDuration="1m36.861563822s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:33.859026751 +0000 UTC m=+117.891638293" watchObservedRunningTime="2026-03-17 18:49:33.861563822 +0000 UTC m=+117.894175344" Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.868981 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" event={"ID":"9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd","Type":"ContainerStarted","Data":"4621d8e72dd2b3b335e6c5a00f184eadbd3e2233a1cb0c92a947b8938e10e48a"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.879258 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" event={"ID":"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952","Type":"ContainerStarted","Data":"9fbb97a6d13018fb55e272b31500d1cfb5c3e2e0e1d49c22d7497f7e2e816810"} Mar 17 18:49:33 crc kubenswrapper[5110]: I0317 18:49:33.883856 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-f8wb2" podStartSLOduration=6.883837091 podStartE2EDuration="6.883837091s" podCreationTimestamp="2026-03-17 18:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:33.879544862 +0000 UTC m=+117.912156384" watchObservedRunningTime="2026-03-17 18:49:33.883837091 +0000 UTC m=+117.916448613" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:33.980248 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:33.980525 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.48051113 +0000 UTC m=+118.513122652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.042227 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" event={"ID":"deb743a6-6308-466e-830a-ae0a66472896","Type":"ContainerStarted","Data":"02bb7bc11efc8d368a3c3bcb0e7701d300d2cdb6e959c734686b31ae28d0f83f"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.067756 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-74545575db-lz8zr" event={"ID":"8477e2b6-663f-49b8-a714-5925de307ac9","Type":"ContainerStarted","Data":"35537dcd02b5298c62c54d489135f59a02b782290d73584136959fd4bf3eb9be"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.080400 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" event={"ID":"54a6c559-52e9-4f87-88f6-0d0ab99aeebc","Type":"ContainerStarted","Data":"de9348e8dbde6f1b6e46520e2c3ae0b855d49afb649c3634566753132324d773"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.081766 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.082541 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.582524648 +0000 UTC m=+118.615136160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.195625 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.196917 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.69689791 +0000 UTC m=+118.729509432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.210297 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" event={"ID":"f8606fbd-7611-4970-9f65-f3e12bb76894","Type":"ContainerStarted","Data":"43240751477aa54fda9f7eb65864eb0a1ca5892061068366894d68f2fa21a636"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.218853 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" event={"ID":"69e1d569-06c8-4098-8e3f-57c1a5970756","Type":"ContainerStarted","Data":"bd148730446ecdbc85f4befd1ad3b8d54c0fcf011e11bfcb1e3a9d8da8b48e82"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.220185 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.222522 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" podStartSLOduration=96.222506272 podStartE2EDuration="1m36.222506272s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:33.979724008 +0000 UTC m=+118.012335550" watchObservedRunningTime="2026-03-17 18:49:34.222506272 +0000 UTC m=+118.255117794" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.233139 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" event={"ID":"7930ff4e-b5eb-42d7-ab81-884f18b423d7","Type":"ContainerStarted","Data":"2d939759080a81d7c7a0889573a001683445bc5552233047604729d720ad1b2e"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.240639 5110 patch_prober.go:28] interesting pod/packageserver-7d4fc7d867-k7cdw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" start-of-body= Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.240707 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" podUID="69e1d569-06c8-4098-8e3f-57c1a5970756" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.256282 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" podStartSLOduration=96.256266441 podStartE2EDuration="1m36.256266441s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.255440118 +0000 UTC m=+118.288051650" watchObservedRunningTime="2026-03-17 18:49:34.256266441 +0000 UTC m=+118.288877963" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.257405 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podStartSLOduration=96.257396972 podStartE2EDuration="1m36.257396972s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.234881286 +0000 UTC m=+118.267492818" watchObservedRunningTime="2026-03-17 18:49:34.257396972 +0000 UTC m=+118.290008494" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.262135 5110 generic.go:358] "Generic (PLEG): container finished" podID="ab45c12d-7f9a-4053-b810-645c966906a1" containerID="bf57ec5155cb84b27cd417838f37ec3dc0d8e6f7d3da8721bbc1fae2018a5971" exitCode=0 Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.262278 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" event={"ID":"ab45c12d-7f9a-4053-b810-645c966906a1","Type":"ContainerDied","Data":"bf57ec5155cb84b27cd417838f37ec3dc0d8e6f7d3da8721bbc1fae2018a5971"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.300424 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" event={"ID":"10e07288-52f4-407c-a8eb-709214e358f5","Type":"ContainerStarted","Data":"6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a"} Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.300478 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.302539 5110 patch_prober.go:28] interesting pod/downloads-747b44746d-86ddn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.302897 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-86ddn" podUID="9731f229-c1b8-4511-819c-b90e6cd6af2e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.317324 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.326751 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.826734781 +0000 UTC m=+118.859346303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.378819 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-67c89758df-tnkj2" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.384866 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-74545575db-lz8zr" podStartSLOduration=96.384843637 podStartE2EDuration="1m36.384843637s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.379218221 +0000 UTC m=+118.411829753" watchObservedRunningTime="2026-03-17 18:49:34.384843637 +0000 UTC m=+118.417455159" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.423956 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.424421 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:34.924399948 +0000 UTC m=+118.957011470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.432272 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-8596bd845d-g59zg" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.433473 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" podStartSLOduration=97.43345584 podStartE2EDuration="1m37.43345584s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.432697658 +0000 UTC m=+118.465309180" watchObservedRunningTime="2026-03-17 18:49:34.43345584 +0000 UTC m=+118.466067362" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.438047 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50822: no serving certificate available for the kubelet" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.453196 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" podStartSLOduration=96.453177138 podStartE2EDuration="1m36.453177138s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.452290804 +0000 UTC m=+118.484902346" watchObservedRunningTime="2026-03-17 18:49:34.453177138 +0000 UTC m=+118.485788670" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.458811 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.514885 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-7t4sr" podStartSLOduration=97.514866884 podStartE2EDuration="1m37.514866884s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.473171274 +0000 UTC m=+118.505782806" watchObservedRunningTime="2026-03-17 18:49:34.514866884 +0000 UTC m=+118.547478406" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.525159 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.525443 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.025430838 +0000 UTC m=+119.058042360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.566211 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50830: no serving certificate available for the kubelet" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.573685 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" podStartSLOduration=7.57366411 podStartE2EDuration="7.57366411s" podCreationTimestamp="2026-03-17 18:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:34.570136841 +0000 UTC m=+118.602748353" watchObservedRunningTime="2026-03-17 18:49:34.57366411 +0000 UTC m=+118.606275632" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.641602 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.641898 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.141882517 +0000 UTC m=+119.174494039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.747799 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.748390 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.24837828 +0000 UTC m=+119.280989792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.764293 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50844: no serving certificate available for the kubelet" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.813325 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:34 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:34 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:34 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.813383 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.836082 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jqrnn"] Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.839026 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50854: no serving certificate available for the kubelet" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.849118 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.849477 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.349460261 +0000 UTC m=+119.382071783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.969244 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50858: no serving certificate available for the kubelet" Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.969660 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:34 crc kubenswrapper[5110]: E0317 18:49:34.970064 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.470037865 +0000 UTC m=+119.502649387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:34 crc kubenswrapper[5110]: I0317 18:49:34.971258 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.048627 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50866: no serving certificate available for the kubelet" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.070912 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.071147 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.571116947 +0000 UTC m=+119.603728469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.150169 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50870: no serving certificate available for the kubelet" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.172207 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.172584 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.672572619 +0000 UTC m=+119.705184141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.273717 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50878: no serving certificate available for the kubelet" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.282780 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.283217 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.783200396 +0000 UTC m=+119.815811918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.322019 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" event={"ID":"152126f1-6413-47f3-9847-ca6e3e561c7b","Type":"ContainerStarted","Data":"9269346dd1601b46e7b5728871a82ecad26770114cabd4e6dd3815e36cf9bda4"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.333236 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" event={"ID":"ab45c12d-7f9a-4053-b810-645c966906a1","Type":"ContainerStarted","Data":"167797ebb7b901ec8d2a42ac76f0305d8abf30b8d9f0b47a2b74b08cf1cb0a43"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.338351 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" event={"ID":"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5","Type":"ContainerStarted","Data":"b8e1d8bfcc1bf34ea4250ceb8b99b5a12fcc1246938f9cc7324993f3b33df7d3"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.338559 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.340490 5110 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-95rbr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.340568 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.343137 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" event={"ID":"023fe200-90c4-43c0-aa37-2241d9061e06","Type":"ContainerStarted","Data":"fb31756a00583f83115973ebae16dd680611b8a77535ec202b81785b7155e3b3"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.345586 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-s79f7" podStartSLOduration=97.345571571 podStartE2EDuration="1m37.345571571s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.344132151 +0000 UTC m=+119.376743693" watchObservedRunningTime="2026-03-17 18:49:35.345571571 +0000 UTC m=+119.378183093" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.360603 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" event={"ID":"57ae9010-5764-463a-93ad-d4036ee2b53c","Type":"ContainerStarted","Data":"8815dff18a97e06b0f9c097f8d924082574c440eebb8171d7df956dde429ccf0"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.373875 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" event={"ID":"967f9e08-ced9-444e-b03a-439f0f4cc686","Type":"ContainerStarted","Data":"848f4dfd863f39ee70911f9143bdf96185bb3fed20b4ba3401a31af912fe836b"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.378754 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" event={"ID":"e2337485-1832-41b3-97d8-4ccd31ed60a3","Type":"ContainerStarted","Data":"69b8a82520864ff19ac970c2e1b33b2f6e091a66c317ed93d18213c5016d3cf0"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.379497 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.399772 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.400249 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-gzdmn" podStartSLOduration=97.400235831 podStartE2EDuration="1m37.400235831s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.399342466 +0000 UTC m=+119.431954008" watchObservedRunningTime="2026-03-17 18:49:35.400235831 +0000 UTC m=+119.432847353" Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.400776 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:35.900756886 +0000 UTC m=+119.933368408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.401592 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" podStartSLOduration=97.401583369 podStartE2EDuration="1m37.401583369s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.3707082 +0000 UTC m=+119.403319732" watchObservedRunningTime="2026-03-17 18:49:35.401583369 +0000 UTC m=+119.434194911" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.426295 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" event={"ID":"9b20488e-b2ad-4a1d-83e7-18c7e27ca8bd","Type":"ContainerStarted","Data":"39e51db27872670887c2f93d45a0c7b5b0c7eaf40f16862ca0312cc3a59873fa"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.432399 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" podStartSLOduration=97.432373155 podStartE2EDuration="1m37.432373155s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.424652331 +0000 UTC m=+119.457263873" watchObservedRunningTime="2026-03-17 18:49:35.432373155 +0000 UTC m=+119.464984687" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.447605 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-xsmjk" podStartSLOduration=97.447587339 podStartE2EDuration="1m37.447587339s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.447517437 +0000 UTC m=+119.480128959" watchObservedRunningTime="2026-03-17 18:49:35.447587339 +0000 UTC m=+119.480198861" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.460654 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" event={"ID":"e7e11c8a-b0f2-4f43-bc25-43cb7f57f952","Type":"ContainerStarted","Data":"636190da409b93b802aa7c7d56f7a95206f7fa4eb9d1497b6d5fcb44db3b43d8"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.487305 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" event={"ID":"97c5ac55-d0d1-4075-84c6-3de5c8388a96","Type":"ContainerStarted","Data":"ac23d28d25340918c226d8abb53638341c65d421c0201f6423f156a02a31159a"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.501391 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" event={"ID":"74bf2564-afaa-4967-9901-3fd670b5b991","Type":"ContainerStarted","Data":"6c63d5e8ae09b5430a427321154981aaac2c746c686b5331d717dcea43283705"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.502537 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.510003 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.009980683 +0000 UTC m=+120.042592205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.518588 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-k5fgh" podStartSLOduration=97.518570332 podStartE2EDuration="1m37.518570332s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.483375213 +0000 UTC m=+119.515986735" watchObservedRunningTime="2026-03-17 18:49:35.518570332 +0000 UTC m=+119.551181854" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.549984 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" event={"ID":"84f8e2e3-408f-4873-90e9-f778b20e1a32","Type":"ContainerStarted","Data":"af9ab56ab5b68e4ab25600cc9986ce228823277fa165d643a23b87261a35bf63"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.550293 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.553661 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" event={"ID":"54a6c559-52e9-4f87-88f6-0d0ab99aeebc","Type":"ContainerStarted","Data":"d94e8920303d9462b169bde898d12e5acafb70ca217f3c0fde586c9212d5ea84"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.567722 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-8gfln" podStartSLOduration=97.567672918 podStartE2EDuration="1m37.567672918s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.520856335 +0000 UTC m=+119.553467857" watchObservedRunningTime="2026-03-17 18:49:35.567672918 +0000 UTC m=+119.600284440" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.569239 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-xkvs8" podStartSLOduration=97.5692082 podStartE2EDuration="1m37.5692082s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.567363929 +0000 UTC m=+119.599975451" watchObservedRunningTime="2026-03-17 18:49:35.5692082 +0000 UTC m=+119.601819722" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.577417 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" event={"ID":"eb2106ad-091a-45e1-8e42-7dd01f174c03","Type":"ContainerStarted","Data":"5f99a6a7e6331f7bc8d3be293d5408b8edd187ee519a34e090a6ed382801463d"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.577475 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" event={"ID":"eb2106ad-091a-45e1-8e42-7dd01f174c03","Type":"ContainerStarted","Data":"d925cbe3228335835d3bb0f580609307b05ebc0b11e919ddbca21fb08e9f1dc2"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.604203 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-lhdwb" podStartSLOduration=97.604183353 podStartE2EDuration="1m37.604183353s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.603798032 +0000 UTC m=+119.636409574" watchObservedRunningTime="2026-03-17 18:49:35.604183353 +0000 UTC m=+119.636794875" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.624543 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.625913 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.125890257 +0000 UTC m=+120.158501779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.627389 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" event={"ID":"f8606fbd-7611-4970-9f65-f3e12bb76894","Type":"ContainerStarted","Data":"bf0fb438c3eda441f0783a7088e6e491aa32aade6aa018d10f255540c8d064ee"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.636281 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" podStartSLOduration=98.636253265 podStartE2EDuration="1m38.636253265s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.627406869 +0000 UTC m=+119.660018411" watchObservedRunningTime="2026-03-17 18:49:35.636253265 +0000 UTC m=+119.668864787" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.642120 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dhxm4" event={"ID":"7923cd83-a73e-4a4b-9bdd-b38f7656579b","Type":"ContainerStarted","Data":"1f1823fe3ad725bde1641522d517b5d820e031c5e1de4c903a9b9c1d1282d53f"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.642351 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dhxm4" event={"ID":"7923cd83-a73e-4a4b-9bdd-b38f7656579b","Type":"ContainerStarted","Data":"887f3814cdef1afdfc5d22097b38f908b04297012cb2abb591866d594a24c6c1"} Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.653154 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-gkd64" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.653363 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-h46kj" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.661240 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-5g6pb" podStartSLOduration=97.66122251 podStartE2EDuration="1m37.66122251s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.658636538 +0000 UTC m=+119.691248080" watchObservedRunningTime="2026-03-17 18:49:35.66122251 +0000 UTC m=+119.693834032" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.685155 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-bn9fl" podStartSLOduration=97.685136695 podStartE2EDuration="1m37.685136695s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.681172575 +0000 UTC m=+119.713784097" watchObservedRunningTime="2026-03-17 18:49:35.685136695 +0000 UTC m=+119.717748217" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.739740 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.748546 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.749567 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.249546177 +0000 UTC m=+120.282157699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.750438 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-jx99w" podStartSLOduration=98.750425691 podStartE2EDuration="1m38.750425691s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.745890575 +0000 UTC m=+119.778502097" watchObservedRunningTime="2026-03-17 18:49:35.750425691 +0000 UTC m=+119.783037213" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.808432 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:35 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:35 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:35 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.808517 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.826096 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-69db94689b-z7gg7" podStartSLOduration=97.826082155 podStartE2EDuration="1m37.826082155s" podCreationTimestamp="2026-03-17 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.823486413 +0000 UTC m=+119.856097935" watchObservedRunningTime="2026-03-17 18:49:35.826082155 +0000 UTC m=+119.858693677" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.843008 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.843313 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.343300674 +0000 UTC m=+120.375912196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.855027 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dhxm4" podStartSLOduration=8.854988759 podStartE2EDuration="8.854988759s" podCreationTimestamp="2026-03-17 18:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:35.850879635 +0000 UTC m=+119.883491167" watchObservedRunningTime="2026-03-17 18:49:35.854988759 +0000 UTC m=+119.887600281" Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.944645 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.944851 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.444821168 +0000 UTC m=+120.477432690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.944955 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:35 crc kubenswrapper[5110]: E0317 18:49:35.945292 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.445277631 +0000 UTC m=+120.477889223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:35 crc kubenswrapper[5110]: I0317 18:49:35.994555 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50882: no serving certificate available for the kubelet" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.047014 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.047208 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.547181725 +0000 UTC m=+120.579793257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.047353 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.047685 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.547668509 +0000 UTC m=+120.580280031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.148437 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.148580 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.648553655 +0000 UTC m=+120.681165177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.148774 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.149202 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.649180483 +0000 UTC m=+120.681792005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.250438 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.250950 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.750934163 +0000 UTC m=+120.783545685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.352667 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.352995 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.852979662 +0000 UTC m=+120.885591184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.454179 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.454389 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:36.954357721 +0000 UTC m=+120.986969243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.556712 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.557046 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.057029817 +0000 UTC m=+121.089641339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.596387 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k7cdw" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.658761 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.658948 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.158918381 +0000 UTC m=+121.191529903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.659486 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.659837 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.159820987 +0000 UTC m=+121.192432509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.682244 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.699847 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48"} Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.701076 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.763021 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.763737 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.263703666 +0000 UTC m=+121.296315198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.787843 5110 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-95rbr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.787971 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" gracePeriod=30 Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.788433 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.794070 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.794480 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.294462632 +0000 UTC m=+121.327074154 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.816226 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:36 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:36 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:36 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.816293 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.819477 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=25.819458207 podStartE2EDuration="25.819458207s" podCreationTimestamp="2026-03-17 18:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:36.8166984 +0000 UTC m=+120.849309922" watchObservedRunningTime="2026-03-17 18:49:36.819458207 +0000 UTC m=+120.852069729" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.866588 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.866617 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nldd2"] Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.880962 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nldd2"] Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.881020 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" event={"ID":"ab45c12d-7f9a-4053-b810-645c966906a1","Type":"ContainerStarted","Data":"d4b3c82e8dff1507202bca0fb2faf48076ec9d1633e5ba4f9b965faa2bf7c034"} Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.881073 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-czltc"] Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.896010 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:36 crc kubenswrapper[5110]: E0317 18:49:36.898334 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.39831485 +0000 UTC m=+121.430926372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.908987 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-czltc"] Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.909107 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.909340 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.909768 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" podStartSLOduration=99.909760179 podStartE2EDuration="1m39.909760179s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:36.903750002 +0000 UTC m=+120.936361544" watchObservedRunningTime="2026-03-17 18:49:36.909760179 +0000 UTC m=+120.942371701" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.913316 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Mar 17 18:49:36 crc kubenswrapper[5110]: I0317 18:49:36.913391 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.000371 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4hxl\" (UniqueName: \"kubernetes.io/projected/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-kube-api-access-q4hxl\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.002470 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czp6q\" (UniqueName: \"kubernetes.io/projected/72bdf6fb-83a5-4dcb-996c-0bee47903667-kube-api-access-czp6q\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.002550 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-catalog-content\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.002615 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-utilities\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.002638 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-catalog-content\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.002682 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-utilities\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.003090 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.003387 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.503374033 +0000 UTC m=+121.535985555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.060960 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4vpt2"] Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.069591 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.072887 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4vpt2"] Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.137329 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.137947 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q4hxl\" (UniqueName: \"kubernetes.io/projected/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-kube-api-access-q4hxl\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.137978 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-czp6q\" (UniqueName: \"kubernetes.io/projected/72bdf6fb-83a5-4dcb-996c-0bee47903667-kube-api-access-czp6q\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.138041 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-catalog-content\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.139951 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-catalog-content\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.144345 5110 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7930ff4e_b5eb_42d7_ab81_884f18b423d7.slice/crio-conmon-2d939759080a81d7c7a0889573a001683445bc5552233047604729d720ad1b2e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7930ff4e_b5eb_42d7_ab81_884f18b423d7.slice/crio-2d939759080a81d7c7a0889573a001683445bc5552233047604729d720ad1b2e.scope\": RecentStats: unable to find data in memory cache]" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.154189 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.654146097 +0000 UTC m=+121.686757619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.154331 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-utilities\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.154355 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-catalog-content\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.154402 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-utilities\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.154500 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.154904 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.654896397 +0000 UTC m=+121.687507919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.155450 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-utilities\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.161950 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-utilities\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.168577 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-catalog-content\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.197223 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-czp6q\" (UniqueName: \"kubernetes.io/projected/72bdf6fb-83a5-4dcb-996c-0bee47903667-kube-api-access-czp6q\") pod \"certified-operators-nldd2\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.227796 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4hxl\" (UniqueName: \"kubernetes.io/projected/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-kube-api-access-q4hxl\") pod \"community-operators-czltc\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.266552 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.266903 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-775bm\" (UniqueName: \"kubernetes.io/projected/5f27c690-db4e-404f-88d9-60f93687021d-kube-api-access-775bm\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.266962 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-catalog-content\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.267007 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-utilities\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.267208 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.767187081 +0000 UTC m=+121.799798613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.267455 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.271293 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.310048 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rx699"] Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.370612 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-775bm\" (UniqueName: \"kubernetes.io/projected/5f27c690-db4e-404f-88d9-60f93687021d-kube-api-access-775bm\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.370952 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-catalog-content\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.370979 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.371006 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-utilities\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.371399 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-utilities\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.371883 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.871864842 +0000 UTC m=+121.904476364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.371884 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-catalog-content\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.472747 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.473210 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:37.973190431 +0000 UTC m=+122.005801963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.501532 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rx699"] Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.501698 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.512600 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-775bm\" (UniqueName: \"kubernetes.io/projected/5f27c690-db4e-404f-88d9-60f93687021d-kube-api-access-775bm\") pod \"certified-operators-4vpt2\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.527596 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50896: no serving certificate available for the kubelet" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.574951 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.575007 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gjt2\" (UniqueName: \"kubernetes.io/projected/c15d66c3-0baf-44b5-aa64-ff26d681badd-kube-api-access-9gjt2\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.575041 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-utilities\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.575128 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-catalog-content\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.575436 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.075424305 +0000 UTC m=+122.108035827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.683843 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.684087 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9gjt2\" (UniqueName: \"kubernetes.io/projected/c15d66c3-0baf-44b5-aa64-ff26d681badd-kube-api-access-9gjt2\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.684137 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-utilities\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.684236 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-catalog-content\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.684763 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-catalog-content\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.685154 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.185134106 +0000 UTC m=+122.217745638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.686046 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-utilities\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.731023 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gjt2\" (UniqueName: \"kubernetes.io/projected/c15d66c3-0baf-44b5-aa64-ff26d681badd-kube-api-access-9gjt2\") pod \"community-operators-rx699\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.734324 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.787370 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" event={"ID":"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888","Type":"ContainerStarted","Data":"74e4a8fd02ea44bcb6b2810a2cc4753eaf06f9d7f050043089b417d7cd239cf6"} Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.788384 5110 generic.go:358] "Generic (PLEG): container finished" podID="7930ff4e-b5eb-42d7-ab81-884f18b423d7" containerID="2d939759080a81d7c7a0889573a001683445bc5552233047604729d720ad1b2e" exitCode=0 Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.789128 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" event={"ID":"7930ff4e-b5eb-42d7-ab81-884f18b423d7","Type":"ContainerDied","Data":"2d939759080a81d7c7a0889573a001683445bc5552233047604729d720ad1b2e"} Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.806818 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.809915 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.309901137 +0000 UTC m=+122.342512659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.832361 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:37 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:37 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:37 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.832425 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.896003 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:49:37 crc kubenswrapper[5110]: I0317 18:49:37.908500 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:37 crc kubenswrapper[5110]: E0317 18:49:37.908907 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.40888939 +0000 UTC m=+122.441500912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.005601 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nldd2"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.010912 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.011234 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.511219417 +0000 UTC m=+122.543830939 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.055025 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-czltc"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.112013 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.112165 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.612144624 +0000 UTC m=+122.644756146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.112385 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.112730 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.61272354 +0000 UTC m=+122.645335062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.213717 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.213886 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.713859343 +0000 UTC m=+122.746470865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.214174 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.214529 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.714520432 +0000 UTC m=+122.747131954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.317770 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.318126 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.818109943 +0000 UTC m=+122.850721465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.335115 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4vpt2"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.421831 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.422243 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:38.922230629 +0000 UTC m=+122.954842151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.523403 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.523805 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.023785744 +0000 UTC m=+123.056397266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.523897 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.524176 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.024169145 +0000 UTC m=+123.056780657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.622924 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rx699"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.627497 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.627928 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.12791187 +0000 UTC m=+123.160523392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.653399 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bh5rg"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.659987 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.662992 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh5rg"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.663448 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.728789 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-utilities\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.728835 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-catalog-content\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.728988 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.729205 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gr2k\" (UniqueName: \"kubernetes.io/projected/9fd9df75-65d5-476e-9e82-da35f1116996-kube-api-access-9gr2k\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.729553 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.229538647 +0000 UTC m=+123.262150169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: W0317 18:49:38.768521 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc15d66c3_0baf_44b5_aa64_ff26d681badd.slice/crio-8ec46c9c115dcdd6b68e0b2e35eaa2fa87336d0be3a444fcd9106f07252b51c3 WatchSource:0}: Error finding container 8ec46c9c115dcdd6b68e0b2e35eaa2fa87336d0be3a444fcd9106f07252b51c3: Status 404 returned error can't find the container with id 8ec46c9c115dcdd6b68e0b2e35eaa2fa87336d0be3a444fcd9106f07252b51c3 Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.802117 5110 generic.go:358] "Generic (PLEG): container finished" podID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerID="ae938c8d26b6fc6a3dc1dc7c2ca94d7be64cc69517a79ec75da05d44542ea2c7" exitCode=0 Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.802248 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czltc" event={"ID":"a0b63bb8-f8c1-4591-affd-65e1d6cb3393","Type":"ContainerDied","Data":"ae938c8d26b6fc6a3dc1dc7c2ca94d7be64cc69517a79ec75da05d44542ea2c7"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.802279 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czltc" event={"ID":"a0b63bb8-f8c1-4591-affd-65e1d6cb3393","Type":"ContainerStarted","Data":"13ddd43aea779eaf69faa20c0ea571ded8a9fa997b9cd5977b2a945417f7cdf5"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.804849 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:38 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:38 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:38 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.804894 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.809622 5110 generic.go:358] "Generic (PLEG): container finished" podID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerID="be05241bad40425d9a516f897f859b0220916a46d99d78693d71dd29ba1aa337" exitCode=0 Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.810100 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldd2" event={"ID":"72bdf6fb-83a5-4dcb-996c-0bee47903667","Type":"ContainerDied","Data":"be05241bad40425d9a516f897f859b0220916a46d99d78693d71dd29ba1aa337"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.810131 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldd2" event={"ID":"72bdf6fb-83a5-4dcb-996c-0bee47903667","Type":"ContainerStarted","Data":"8bc77024ac1a104eadb739838a88ed32b8d771568727e3c40767b594bdb41554"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.831094 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.831270 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.331243836 +0000 UTC m=+123.363855358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.831630 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-utilities\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.831680 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-catalog-content\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.832017 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.832360 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-catalog-content\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.832456 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.33244097 +0000 UTC m=+123.365052492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.833346 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9gr2k\" (UniqueName: \"kubernetes.io/projected/9fd9df75-65d5-476e-9e82-da35f1116996-kube-api-access-9gr2k\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.833432 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-utilities\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.843404 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx699" event={"ID":"c15d66c3-0baf-44b5-aa64-ff26d681badd","Type":"ContainerStarted","Data":"8ec46c9c115dcdd6b68e0b2e35eaa2fa87336d0be3a444fcd9106f07252b51c3"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.861314 5110 generic.go:358] "Generic (PLEG): container finished" podID="5f27c690-db4e-404f-88d9-60f93687021d" containerID="ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d" exitCode=0 Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.862719 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerDied","Data":"ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.862747 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerStarted","Data":"3a136d18084e4722f55612ff0d1cfed91420e9b4216935a0c98d27e67fddae7a"} Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.873992 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gr2k\" (UniqueName: \"kubernetes.io/projected/9fd9df75-65d5-476e-9e82-da35f1116996-kube-api-access-9gr2k\") pod \"redhat-marketplace-bh5rg\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.934928 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.935093 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.435070124 +0000 UTC m=+123.467681646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.935279 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:38 crc kubenswrapper[5110]: E0317 18:49:38.935616 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.435606429 +0000 UTC m=+123.468217951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.965933 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.988176 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.994939 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler\"/\"installer-sa-dockercfg-qpkss\"" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.995253 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler\"/\"kube-root-ca.crt\"" Mar 17 18:49:38 crc kubenswrapper[5110]: I0317 18:49:38.997559 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.037676 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.037919 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.537852803 +0000 UTC m=+123.570464325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.038204 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.038262 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.038408 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.038683 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.538674546 +0000 UTC m=+123.571286068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.066545 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zw8tx"] Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.071039 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.082515 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.100983 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zw8tx"] Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.152660 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.152845 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-catalog-content\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.152962 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-utilities\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.153033 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.153071 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kshp\" (UniqueName: \"kubernetes.io/projected/ba6a6e05-30d4-4581-95ac-3ca772fad52e-kube-api-access-2kshp\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.153110 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.153229 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.653212111 +0000 UTC m=+123.685823633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.153273 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.185870 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.230694 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.253958 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7930ff4e-b5eb-42d7-ab81-884f18b423d7-config-volume\") pod \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.254156 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srk6n\" (UniqueName: \"kubernetes.io/projected/7930ff4e-b5eb-42d7-ab81-884f18b423d7-kube-api-access-srk6n\") pod \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.254220 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7930ff4e-b5eb-42d7-ab81-884f18b423d7-secret-volume\") pod \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\" (UID: \"7930ff4e-b5eb-42d7-ab81-884f18b423d7\") " Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.254328 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-catalog-content\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.254378 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.254441 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-utilities\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.254478 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2kshp\" (UniqueName: \"kubernetes.io/projected/ba6a6e05-30d4-4581-95ac-3ca772fad52e-kube-api-access-2kshp\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.256637 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-catalog-content\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.256913 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.756896115 +0000 UTC m=+123.789507637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.257307 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-utilities\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.291811 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7930ff4e-b5eb-42d7-ab81-884f18b423d7-config-volume" (OuterVolumeSpecName: "config-volume") pod "7930ff4e-b5eb-42d7-ab81-884f18b423d7" (UID: "7930ff4e-b5eb-42d7-ab81-884f18b423d7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.294475 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kshp\" (UniqueName: \"kubernetes.io/projected/ba6a6e05-30d4-4581-95ac-3ca772fad52e-kube-api-access-2kshp\") pod \"redhat-marketplace-zw8tx\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.296074 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7930ff4e-b5eb-42d7-ab81-884f18b423d7-kube-api-access-srk6n" (OuterVolumeSpecName: "kube-api-access-srk6n") pod "7930ff4e-b5eb-42d7-ab81-884f18b423d7" (UID: "7930ff4e-b5eb-42d7-ab81-884f18b423d7"). InnerVolumeSpecName "kube-api-access-srk6n". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.310146 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7930ff4e-b5eb-42d7-ab81-884f18b423d7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7930ff4e-b5eb-42d7-ab81-884f18b423d7" (UID: "7930ff4e-b5eb-42d7-ab81-884f18b423d7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.335670 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.356174 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.356334 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.8563025 +0000 UTC m=+123.888914032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.356567 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.356780 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-srk6n\" (UniqueName: \"kubernetes.io/projected/7930ff4e-b5eb-42d7-ab81-884f18b423d7-kube-api-access-srk6n\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.356802 5110 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7930ff4e-b5eb-42d7-ab81-884f18b423d7-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.356814 5110 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7930ff4e-b5eb-42d7-ab81-884f18b423d7-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.357022 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.85700784 +0000 UTC m=+123.889619362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.427525 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.458263 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.458769 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:39.95875001 +0000 UTC m=+123.991361532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.491741 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh5rg"] Mar 17 18:49:39 crc kubenswrapper[5110]: W0317 18:49:39.504017 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fd9df75_65d5_476e_9e82_da35f1116996.slice/crio-ee503274009754b42e63290144f3ac7ab5adc668a60a9feeae54ec857a788f04 WatchSource:0}: Error finding container ee503274009754b42e63290144f3ac7ab5adc668a60a9feeae54ec857a788f04: Status 404 returned error can't find the container with id ee503274009754b42e63290144f3ac7ab5adc668a60a9feeae54ec857a788f04 Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.560145 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.560495 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.06047902 +0000 UTC m=+124.093090542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.661710 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.662000 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.161977383 +0000 UTC m=+124.194588905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.702722 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Mar 17 18:49:39 crc kubenswrapper[5110]: W0317 18:49:39.718270 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4da4da15_d71d_40e2_b44b_54f0fbed3b6c.slice/crio-535006442a9c882b822d58b3de75a72636ff4d9685c6d2aeffb8e36211e8a17e WatchSource:0}: Error finding container 535006442a9c882b822d58b3de75a72636ff4d9685c6d2aeffb8e36211e8a17e: Status 404 returned error can't find the container with id 535006442a9c882b822d58b3de75a72636ff4d9685c6d2aeffb8e36211e8a17e Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.752617 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zw8tx"] Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.766493 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.767267 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.266941943 +0000 UTC m=+124.299553455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.807265 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:39 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:39 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:39 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.807318 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.808514 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-5777786469-h787x" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.868825 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.869246 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.369217657 +0000 UTC m=+124.401829179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.887219 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" event={"ID":"7930ff4e-b5eb-42d7-ab81-884f18b423d7","Type":"ContainerDied","Data":"cc8bf6b8dda570da86086544d743910903afd53a89a3a04cc08313eac3def8dd"} Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.887261 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc8bf6b8dda570da86086544d743910903afd53a89a3a04cc08313eac3def8dd" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.887271 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562885-nwdmv" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.895048 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jctgs"] Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.896324 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7930ff4e-b5eb-42d7-ab81-884f18b423d7" containerName="collect-profiles" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.896346 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7930ff4e-b5eb-42d7-ab81-884f18b423d7" containerName="collect-profiles" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.896485 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="7930ff4e-b5eb-42d7-ab81-884f18b423d7" containerName="collect-profiles" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.899472 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.900214 5110 generic.go:358] "Generic (PLEG): container finished" podID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerID="502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4" exitCode=0 Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.901402 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx699" event={"ID":"c15d66c3-0baf-44b5-aa64-ff26d681badd","Type":"ContainerDied","Data":"502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4"} Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.901557 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.907271 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"4da4da15-d71d-40e2-b44b-54f0fbed3b6c","Type":"ContainerStarted","Data":"535006442a9c882b822d58b3de75a72636ff4d9685c6d2aeffb8e36211e8a17e"} Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.909169 5110 generic.go:358] "Generic (PLEG): container finished" podID="9fd9df75-65d5-476e-9e82-da35f1116996" containerID="79088d6b5a1462e75a1de3a8d7242f8f4ba59110b5dbda757fba96e3f8771a70" exitCode=0 Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.909243 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh5rg" event={"ID":"9fd9df75-65d5-476e-9e82-da35f1116996","Type":"ContainerDied","Data":"79088d6b5a1462e75a1de3a8d7242f8f4ba59110b5dbda757fba96e3f8771a70"} Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.909267 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh5rg" event={"ID":"9fd9df75-65d5-476e-9e82-da35f1116996","Type":"ContainerStarted","Data":"ee503274009754b42e63290144f3ac7ab5adc668a60a9feeae54ec857a788f04"} Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.916567 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jctgs"] Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.922212 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerStarted","Data":"77dce939691c0e8e2393fc89a96d0251f64e27b0c697d82e1f8506e049773e20"} Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.970257 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.970305 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rt98\" (UniqueName: \"kubernetes.io/projected/66594344-3d62-414a-9e89-3ab843c5239a-kube-api-access-8rt98\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.970397 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-utilities\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:39 crc kubenswrapper[5110]: I0317 18:49:39.970571 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-catalog-content\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:39 crc kubenswrapper[5110]: E0317 18:49:39.970919 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.470903746 +0000 UTC m=+124.503515268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.072371 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.072608 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-utilities\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.072704 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-catalog-content\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.073004 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8rt98\" (UniqueName: \"kubernetes.io/projected/66594344-3d62-414a-9e89-3ab843c5239a-kube-api-access-8rt98\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.073606 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-utilities\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.073708 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.573687975 +0000 UTC m=+124.606299497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.073926 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-catalog-content\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.093890 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rt98\" (UniqueName: \"kubernetes.io/projected/66594344-3d62-414a-9e89-3ab843c5239a-kube-api-access-8rt98\") pod \"redhat-operators-jctgs\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.123538 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50912: no serving certificate available for the kubelet" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.173801 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.174131 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.674119368 +0000 UTC m=+124.706730890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.239111 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.241365 5110 patch_prober.go:28] interesting pod/downloads-747b44746d-86ddn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.241420 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-747b44746d-86ddn" podUID="9731f229-c1b8-4511-819c-b90e6cd6af2e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.248429 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-955md"] Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.256187 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.263865 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-955md"] Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.276952 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.277093 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.277351 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.277524 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.777506844 +0000 UTC m=+124.810118366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.277677 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxm47\" (UniqueName: \"kubernetes.io/projected/2b3686e6-f0ef-4197-867d-67617aff597e-kube-api-access-rxm47\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.277836 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.277937 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-catalog-content\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.277986 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-utilities\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.278990 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.778979565 +0000 UTC m=+124.811591077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.286016 5110 patch_prober.go:28] interesting pod/apiserver-9ddfb9f55-wcbht container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]log ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]etcd ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/generic-apiserver-start-informers ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/max-in-flight-filter ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 17 18:49:40 crc kubenswrapper[5110]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 17 18:49:40 crc kubenswrapper[5110]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/project.openshift.io-projectcache ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-startinformers ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 17 18:49:40 crc kubenswrapper[5110]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 17 18:49:40 crc kubenswrapper[5110]: livez check failed Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.286106 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" podUID="ab45c12d-7f9a-4053-b810-645c966906a1" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.292028 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.292104 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.307813 5110 patch_prober.go:28] interesting pod/console-64d44f6ddf-8jwk2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.307893 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-console/console-64d44f6ddf-8jwk2" podUID="23f9945f-bddd-4c0a-bf2c-cc0baee81974" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.379233 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.379804 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.879780369 +0000 UTC m=+124.912391891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.380250 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rxm47\" (UniqueName: \"kubernetes.io/projected/2b3686e6-f0ef-4197-867d-67617aff597e-kube-api-access-rxm47\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.380409 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.380834 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-catalog-content\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.380912 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-utilities\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.381403 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.881385184 +0000 UTC m=+124.913996776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.381474 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-catalog-content\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.381913 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-utilities\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.404369 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxm47\" (UniqueName: \"kubernetes.io/projected/2b3686e6-f0ef-4197-867d-67617aff597e-kube-api-access-rxm47\") pod \"redhat-operators-955md\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.482647 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.482970 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:40.982953419 +0000 UTC m=+125.015564941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.584159 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.584505 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.084488343 +0000 UTC m=+125.117099865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.607859 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.676833 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jctgs"] Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.685405 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.685523 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.185502893 +0000 UTC m=+125.218114415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.685670 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.685969 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.185961646 +0000 UTC m=+125.218573168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.704209 5110 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 17 18:49:40 crc kubenswrapper[5110]: W0317 18:49:40.724194 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66594344_3d62_414a_9e89_3ab843c5239a.slice/crio-7dd9f9bea04bda9234d2c6ee3bcc6614399452cbdd04c713668449b94d711a0a WatchSource:0}: Error finding container 7dd9f9bea04bda9234d2c6ee3bcc6614399452cbdd04c713668449b94d711a0a: Status 404 returned error can't find the container with id 7dd9f9bea04bda9234d2c6ee3bcc6614399452cbdd04c713668449b94d711a0a Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.789980 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.790722 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.290694729 +0000 UTC m=+125.323306261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.801824 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.807646 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:40 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:40 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:40 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.807737 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.891413 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.891933 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.391893204 +0000 UTC m=+125.424504726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.928904 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-955md"] Mar 17 18:49:40 crc kubenswrapper[5110]: W0317 18:49:40.950119 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b3686e6_f0ef_4197_867d_67617aff597e.slice/crio-af911c1b82538ea5188357c043eb789df4f0389cb9f41dea59cf8decdea10e4e WatchSource:0}: Error finding container af911c1b82538ea5188357c043eb789df4f0389cb9f41dea59cf8decdea10e4e: Status 404 returned error can't find the container with id af911c1b82538ea5188357c043eb789df4f0389cb9f41dea59cf8decdea10e4e Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.952871 5110 generic.go:358] "Generic (PLEG): container finished" podID="4da4da15-d71d-40e2-b44b-54f0fbed3b6c" containerID="ff0d03f1ecbbab845b516fdd144ba322460b388171df3032dd464a736baceaa2" exitCode=0 Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.953401 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"4da4da15-d71d-40e2-b44b-54f0fbed3b6c","Type":"ContainerDied","Data":"ff0d03f1ecbbab845b516fdd144ba322460b388171df3032dd464a736baceaa2"} Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.957154 5110 generic.go:358] "Generic (PLEG): container finished" podID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerID="14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816" exitCode=0 Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.957263 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerDied","Data":"14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816"} Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.960298 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerStarted","Data":"7dd9f9bea04bda9234d2c6ee3bcc6614399452cbdd04c713668449b94d711a0a"} Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.964517 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" event={"ID":"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888","Type":"ContainerStarted","Data":"e5cc91394b529c8c0c783978856f824f3a20e79ee15a4cc86e4fc7e4abb42db2"} Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.992690 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.992916 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.492879033 +0000 UTC m=+125.525490555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:40 crc kubenswrapper[5110]: I0317 18:49:40.993370 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:40 crc kubenswrapper[5110]: E0317 18:49:40.993885 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.493874881 +0000 UTC m=+125.526486413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.095136 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.095346 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.595321762 +0000 UTC m=+125.627933294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.095647 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.095966 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.59595581 +0000 UTC m=+125.628567342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.196667 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.196797 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.696778125 +0000 UTC m=+125.729389647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.197328 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.197603 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.697595957 +0000 UTC m=+125.730207479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.299325 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.299434 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.799409979 +0000 UTC m=+125.832021501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.299871 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.300222 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.800211592 +0000 UTC m=+125.832823114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.401464 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.401614 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.901590932 +0000 UTC m=+125.934202454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.401916 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: E0317 18:49:41.402371 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-17 18:49:41.902358443 +0000 UTC m=+125.934969975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-5hkdz" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.443657 5110 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-17T18:49:40.704234854Z","UUID":"a40058d5-dc64-4278-8f20-4bd4525ee001","Handler":null,"Name":"","Endpoint":""} Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.450478 5110 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.450517 5110 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.502770 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.513161 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (OuterVolumeSpecName: "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2". PluginName "kubernetes.io/csi", VolumeGIDValue "" Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.604216 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.608080 5110 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.608119 5110 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1264ac67579ad07e7e9003054d44fe40dd55285a4b2f7dc74e48be1aee0868a/globalmount\"" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.652799 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-5hkdz\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.798557 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.804451 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:41 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:41 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:41 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.804525 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:41 crc kubenswrapper[5110]: I0317 18:49:41.808034 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.027827 5110 generic.go:358] "Generic (PLEG): container finished" podID="2b3686e6-f0ef-4197-867d-67617aff597e" containerID="24bb633ad12b8741d94fa92085e649159aebcc5d4eafc288a70c7dbf83add4c6" exitCode=0 Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.028225 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerDied","Data":"24bb633ad12b8741d94fa92085e649159aebcc5d4eafc288a70c7dbf83add4c6"} Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.028254 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerStarted","Data":"af911c1b82538ea5188357c043eb789df4f0389cb9f41dea59cf8decdea10e4e"} Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.042698 5110 generic.go:358] "Generic (PLEG): container finished" podID="66594344-3d62-414a-9e89-3ab843c5239a" containerID="b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8" exitCode=0 Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.043043 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerDied","Data":"b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8"} Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.081558 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" event={"ID":"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888","Type":"ContainerStarted","Data":"203688c3f494e1b96a76e95076c5110e1d0c680efdf93583ec9ee56189e71a24"} Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.081617 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" event={"ID":"779e1c5c-8a98-4f8d-91f5-3cdf72ac1888","Type":"ContainerStarted","Data":"2a604811c8916cea0f80a7c5a3288bcb0dab00a60dcc031da4492b211734da5f"} Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.112267 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ddn25" podStartSLOduration=15.112248139 podStartE2EDuration="15.112248139s" podCreationTimestamp="2026-03-17 18:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:42.111738135 +0000 UTC m=+126.144349657" watchObservedRunningTime="2026-03-17 18:49:42.112248139 +0000 UTC m=+126.144859661" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.208849 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-5hkdz"] Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.334792 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.420578 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kubelet-dir\") pod \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.420707 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4da4da15-d71d-40e2-b44b-54f0fbed3b6c" (UID: "4da4da15-d71d-40e2-b44b-54f0fbed3b6c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.421067 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kube-api-access\") pod \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\" (UID: \"4da4da15-d71d-40e2-b44b-54f0fbed3b6c\") " Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.421322 5110 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.433327 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4da4da15-d71d-40e2-b44b-54f0fbed3b6c" (UID: "4da4da15-d71d-40e2-b44b-54f0fbed3b6c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.522733 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4da4da15-d71d-40e2-b44b-54f0fbed3b6c-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.754379 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9b5059-1b3e-4067-a63d-2952cbe863af" path="/var/lib/kubelet/pods/9e9b5059-1b3e-4067-a63d-2952cbe863af/volumes" Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.805395 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:42 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:42 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:42 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:42 crc kubenswrapper[5110]: I0317 18:49:42.805553 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.092185 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" event={"ID":"67450a68-b2ce-4acb-ba06-f4e60fcf28b4","Type":"ContainerStarted","Data":"75579bcc7f6b04e5e922a381669b2715f243ef1666a95d2bded26a21b8a6a387"} Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.092263 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.092274 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" event={"ID":"67450a68-b2ce-4acb-ba06-f4e60fcf28b4","Type":"ContainerStarted","Data":"c32fda45b100eedbdc0125f07d0e7e348e200caccee6a6ffc4c0f35c94322e2a"} Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.095075 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.095090 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"4da4da15-d71d-40e2-b44b-54f0fbed3b6c","Type":"ContainerDied","Data":"535006442a9c882b822d58b3de75a72636ff4d9685c6d2aeffb8e36211e8a17e"} Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.095160 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="535006442a9c882b822d58b3de75a72636ff4d9685c6d2aeffb8e36211e8a17e" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.120444 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" podStartSLOduration=106.12038217 podStartE2EDuration="1m46.12038217s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:49:43.112763418 +0000 UTC m=+127.145374950" watchObservedRunningTime="2026-03-17 18:49:43.12038217 +0000 UTC m=+127.152993692" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.640344 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.640395 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.640429 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.640458 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.643617 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.643691 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.645133 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.647630 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.650360 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4da4da15-d71d-40e2-b44b-54f0fbed3b6c" containerName="pruner" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.650382 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="4da4da15-d71d-40e2-b44b-54f0fbed3b6c" containerName="pruner" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.650519 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="4da4da15-d71d-40e2-b44b-54f0fbed3b6c" containerName="pruner" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.654597 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.658928 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.664738 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.666272 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.741753 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.743218 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.743393 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.762514 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baee355c-9055-42f1-87c4-3dc7a08b68f6-metrics-certs\") pod \"network-metrics-daemon-m5g58\" (UID: \"baee355c-9055-42f1-87c4-3dc7a08b68f6\") " pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.805844 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:43 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:43 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:43 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.805905 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.818525 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.825663 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.827602 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m5g58" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.831818 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.839469 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.930149 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.930379 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.933095 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Mar 17 18:49:43 crc kubenswrapper[5110]: I0317 18:49:43.933440 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.047490 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.047833 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.156352 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.156453 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.156485 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.180624 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.199339 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m5g58"] Mar 17 18:49:44 crc kubenswrapper[5110]: W0317 18:49:44.219272 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaee355c_9055_42f1_87c4_3dc7a08b68f6.slice/crio-f505a4d68674fb6b4ca4fb95d08bcad93a5353b22f2e4b7c325c78ee977788e3 WatchSource:0}: Error finding container f505a4d68674fb6b4ca4fb95d08bcad93a5353b22f2e4b7c325c78ee977788e3: Status 404 returned error can't find the container with id f505a4d68674fb6b4ca4fb95d08bcad93a5353b22f2e4b7c325c78ee977788e3 Mar 17 18:49:44 crc kubenswrapper[5110]: W0317 18:49:44.262870 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a9ae5f6_97bd_46ac_bafa_ca1b4452a141.slice/crio-f04e545a709d4d7254e7fa3c4ed7bc56df22fad3a35cfc78f61393b9b0be7037 WatchSource:0}: Error finding container f04e545a709d4d7254e7fa3c4ed7bc56df22fad3a35cfc78f61393b9b0be7037: Status 404 returned error can't find the container with id f04e545a709d4d7254e7fa3c4ed7bc56df22fad3a35cfc78f61393b9b0be7037 Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.301494 5110 patch_prober.go:28] interesting pod/downloads-747b44746d-86ddn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.301550 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-86ddn" podUID="9731f229-c1b8-4511-819c-b90e6cd6af2e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 17 18:49:44 crc kubenswrapper[5110]: E0317 18:49:44.302803 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:49:44 crc kubenswrapper[5110]: E0317 18:49:44.304570 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:49:44 crc kubenswrapper[5110]: E0317 18:49:44.306724 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:49:44 crc kubenswrapper[5110]: E0317 18:49:44.306792 5110 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.310773 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.514651 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.807142 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:44 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:44 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:44 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:44 crc kubenswrapper[5110]: I0317 18:49:44.807260 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.111853 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m5g58" event={"ID":"baee355c-9055-42f1-87c4-3dc7a08b68f6","Type":"ContainerStarted","Data":"f505a4d68674fb6b4ca4fb95d08bcad93a5353b22f2e4b7c325c78ee977788e3"} Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.114686 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" event={"ID":"f863fff9-286a-45fa-b8f0-8a86994b8440","Type":"ContainerStarted","Data":"ba55c220e69733d9cf3f84d74dfdc967200d2268b230eabad0daab33b7e735af"} Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.114717 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" event={"ID":"f863fff9-286a-45fa-b8f0-8a86994b8440","Type":"ContainerStarted","Data":"16e3f3784282931b37c624ff5eb2c187112e4cbe9d818e16d50581cb8a44ac3f"} Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.118206 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" event={"ID":"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141","Type":"ContainerStarted","Data":"f04e545a709d4d7254e7fa3c4ed7bc56df22fad3a35cfc78f61393b9b0be7037"} Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.128827 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" event={"ID":"17b87002-b798-480a-8e17-83053d698239","Type":"ContainerStarted","Data":"d90656633d6d17f64eff14531713f995ab912289fc2dd25581c3390a4b99527f"} Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.128870 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" event={"ID":"17b87002-b798-480a-8e17-83053d698239","Type":"ContainerStarted","Data":"24fec5fa56d1c1998103f934880a31e9ff167a4b40629a84632e2c985069cae4"} Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.129480 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.264893 5110 ???:1] "http: TLS handshake error from 192.168.126.11:52994: no serving certificate available for the kubelet" Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.279936 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.292242 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-9ddfb9f55-wcbht" Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.803741 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:45 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:45 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:45 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:45 crc kubenswrapper[5110]: I0317 18:49:45.803813 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:46 crc kubenswrapper[5110]: I0317 18:49:46.789301 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:49:46 crc kubenswrapper[5110]: I0317 18:49:46.793517 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dhxm4" Mar 17 18:49:46 crc kubenswrapper[5110]: I0317 18:49:46.805224 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:46 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:46 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:46 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:46 crc kubenswrapper[5110]: I0317 18:49:46.805299 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:47 crc kubenswrapper[5110]: I0317 18:49:47.142730 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8","Type":"ContainerStarted","Data":"802665a648d6c611942fe87141f3610275a3adedc744e489d00463c47e992942"} Mar 17 18:49:47 crc kubenswrapper[5110]: I0317 18:49:47.796016 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:49:47 crc kubenswrapper[5110]: I0317 18:49:47.805167 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:47 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:47 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:47 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:47 crc kubenswrapper[5110]: I0317 18:49:47.805227 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:48 crc kubenswrapper[5110]: I0317 18:49:48.804221 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:48 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:48 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:48 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:48 crc kubenswrapper[5110]: I0317 18:49:48.804307 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:49 crc kubenswrapper[5110]: I0317 18:49:49.803451 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:49 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:49 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:49 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:49 crc kubenswrapper[5110]: I0317 18:49:49.803819 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:50 crc kubenswrapper[5110]: I0317 18:49:50.292739 5110 patch_prober.go:28] interesting pod/console-64d44f6ddf-8jwk2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 17 18:49:50 crc kubenswrapper[5110]: I0317 18:49:50.292811 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-console/console-64d44f6ddf-8jwk2" podUID="23f9945f-bddd-4c0a-bf2c-cc0baee81974" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 17 18:49:50 crc kubenswrapper[5110]: I0317 18:49:50.805202 5110 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-vhklb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 18:49:50 crc kubenswrapper[5110]: [-]has-synced failed: reason withheld Mar 17 18:49:50 crc kubenswrapper[5110]: [+]process-running ok Mar 17 18:49:50 crc kubenswrapper[5110]: healthz check failed Mar 17 18:49:50 crc kubenswrapper[5110]: I0317 18:49:50.805274 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" podUID="e599baf2-11e5-4d78-9282-b6850af2bd7e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.105986 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-lz6ks"] Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.106299 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerName="controller-manager" containerID="cri-o://54a40f6fcadf19fd5733bfcd79d1e67f1f317709bf00d13f8d23ad9bd8bc8df8" gracePeriod=30 Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.133288 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t"] Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.133551 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" podUID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" containerName="route-controller-manager" containerID="cri-o://bf7af23e9651c1dc674d6207c192895da2ffeee1ce358fb60dfdda02626e311d" gracePeriod=30 Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.305786 5110 patch_prober.go:28] interesting pod/controller-manager-65b6cccf98-lz6ks container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.305843 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.809517 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.812012 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-68cf44c8b8-vhklb" Mar 17 18:49:51 crc kubenswrapper[5110]: I0317 18:49:51.999684 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:49:52 crc kubenswrapper[5110]: I0317 18:49:52.175220 5110 generic.go:358] "Generic (PLEG): container finished" podID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerID="54a40f6fcadf19fd5733bfcd79d1e67f1f317709bf00d13f8d23ad9bd8bc8df8" exitCode=0 Mar 17 18:49:52 crc kubenswrapper[5110]: I0317 18:49:52.175488 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" event={"ID":"85c75701-692d-4bb5-a0d5-2020c652ab82","Type":"ContainerDied","Data":"54a40f6fcadf19fd5733bfcd79d1e67f1f317709bf00d13f8d23ad9bd8bc8df8"} Mar 17 18:49:52 crc kubenswrapper[5110]: I0317 18:49:52.178464 5110 generic.go:358] "Generic (PLEG): container finished" podID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" containerID="bf7af23e9651c1dc674d6207c192895da2ffeee1ce358fb60dfdda02626e311d" exitCode=0 Mar 17 18:49:52 crc kubenswrapper[5110]: I0317 18:49:52.178875 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" event={"ID":"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce","Type":"ContainerDied","Data":"bf7af23e9651c1dc674d6207c192895da2ffeee1ce358fb60dfdda02626e311d"} Mar 17 18:49:54 crc kubenswrapper[5110]: E0317 18:49:54.300981 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:49:54 crc kubenswrapper[5110]: I0317 18:49:54.304577 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-747b44746d-86ddn" Mar 17 18:49:54 crc kubenswrapper[5110]: E0317 18:49:54.306913 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:49:54 crc kubenswrapper[5110]: E0317 18:49:54.308641 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:49:54 crc kubenswrapper[5110]: E0317 18:49:54.308786 5110 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Mar 17 18:49:55 crc kubenswrapper[5110]: I0317 18:49:55.544107 5110 ???:1] "http: TLS handshake error from 192.168.126.11:50374: no serving certificate available for the kubelet" Mar 17 18:49:58 crc kubenswrapper[5110]: I0317 18:49:58.214960 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m5g58" event={"ID":"baee355c-9055-42f1-87c4-3dc7a08b68f6","Type":"ContainerStarted","Data":"f3c6076eb0e7a4070c796bb617108a98fef3ac511ed480144ace26e365ac8722"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.026811 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.050676 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.055791 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6fddb65cd-rtxpl"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.056392 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerName="controller-manager" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.056409 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerName="controller-manager" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.056444 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" containerName="route-controller-manager" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.056450 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" containerName="route-controller-manager" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.056547 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" containerName="controller-manager" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.056557 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" containerName="route-controller-manager" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.065012 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.072474 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fddb65cd-rtxpl"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.103164 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.109177 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.115635 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.133110 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562890-xwcc2"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.149926 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/85c75701-692d-4bb5-a0d5-2020c652ab82-tmp\") pod \"85c75701-692d-4bb5-a0d5-2020c652ab82\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.150471 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c75701-692d-4bb5-a0d5-2020c652ab82-tmp" (OuterVolumeSpecName: "tmp") pod "85c75701-692d-4bb5-a0d5-2020c652ab82" (UID: "85c75701-692d-4bb5-a0d5-2020c652ab82"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.151247 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s9r9\" (UniqueName: \"kubernetes.io/projected/85c75701-692d-4bb5-a0d5-2020c652ab82-kube-api-access-4s9r9\") pod \"85c75701-692d-4bb5-a0d5-2020c652ab82\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.151304 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-client-ca\") pod \"85c75701-692d-4bb5-a0d5-2020c652ab82\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.151363 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-client-ca\") pod \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.151398 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-serving-cert\") pod \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.151459 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-config\") pod \"85c75701-692d-4bb5-a0d5-2020c652ab82\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.151532 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-proxy-ca-bundles\") pod \"85c75701-692d-4bb5-a0d5-2020c652ab82\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.152152 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-client-ca" (OuterVolumeSpecName: "client-ca") pod "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" (UID: "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.152279 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c75701-692d-4bb5-a0d5-2020c652ab82-serving-cert\") pod \"85c75701-692d-4bb5-a0d5-2020c652ab82\" (UID: \"85c75701-692d-4bb5-a0d5-2020c652ab82\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.152400 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-config\") pod \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.152473 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r5j7\" (UniqueName: \"kubernetes.io/projected/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-kube-api-access-8r5j7\") pod \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.152897 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-tmp\") pod \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\" (UID: \"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce\") " Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153233 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da230aff-847c-4f62-8f1c-06f126239d1b-serving-cert\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153381 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-config\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153497 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-config\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153572 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-tmp\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153695 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-client-ca\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.152573 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-config" (OuterVolumeSpecName: "config") pod "85c75701-692d-4bb5-a0d5-2020c652ab82" (UID: "85c75701-692d-4bb5-a0d5-2020c652ab82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153166 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-client-ca" (OuterVolumeSpecName: "client-ca") pod "85c75701-692d-4bb5-a0d5-2020c652ab82" (UID: "85c75701-692d-4bb5-a0d5-2020c652ab82"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153466 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "85c75701-692d-4bb5-a0d5-2020c652ab82" (UID: "85c75701-692d-4bb5-a0d5-2020c652ab82"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153725 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-config" (OuterVolumeSpecName: "config") pod "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" (UID: "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.154151 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-tmp" (OuterVolumeSpecName: "tmp") pod "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" (UID: "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.154165 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153889 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da230aff-847c-4f62-8f1c-06f126239d1b-tmp\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.153531 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562890-xwcc2"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.154901 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-client-ca\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155165 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-proxy-ca-bundles\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155311 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-serving-cert\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155425 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbqhm\" (UniqueName: \"kubernetes.io/projected/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-kube-api-access-jbqhm\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155530 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nktg\" (UniqueName: \"kubernetes.io/projected/da230aff-847c-4f62-8f1c-06f126239d1b-kube-api-access-9nktg\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155672 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155761 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155839 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/85c75701-692d-4bb5-a0d5-2020c652ab82-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.155951 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.156038 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.156147 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.156234 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85c75701-692d-4bb5-a0d5-2020c652ab82-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.156502 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.156912 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.158112 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.159965 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-kube-api-access-8r5j7" (OuterVolumeSpecName: "kube-api-access-8r5j7") pod "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" (UID: "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce"). InnerVolumeSpecName "kube-api-access-8r5j7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.161483 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" (UID: "fc6f0b62-c129-4c6d-a228-af2e48b6a0ce"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.161493 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c75701-692d-4bb5-a0d5-2020c652ab82-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "85c75701-692d-4bb5-a0d5-2020c652ab82" (UID: "85c75701-692d-4bb5-a0d5-2020c652ab82"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.162143 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c75701-692d-4bb5-a0d5-2020c652ab82-kube-api-access-4s9r9" (OuterVolumeSpecName: "kube-api-access-4s9r9") pod "85c75701-692d-4bb5-a0d5-2020c652ab82" (UID: "85c75701-692d-4bb5-a0d5-2020c652ab82"). InnerVolumeSpecName "kube-api-access-4s9r9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.235346 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerStarted","Data":"9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.244951 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerStarted","Data":"2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.255997 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.255996 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" event={"ID":"fc6f0b62-c129-4c6d-a228-af2e48b6a0ce","Type":"ContainerDied","Data":"92da9221935705fd7f5adc4821852df2a489349d90da00979af52c15cd24de13"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.256187 5110 scope.go:117] "RemoveContainer" containerID="bf7af23e9651c1dc674d6207c192895da2ffeee1ce358fb60dfdda02626e311d" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.256963 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-client-ca\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257016 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da230aff-847c-4f62-8f1c-06f126239d1b-tmp\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257088 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-client-ca\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257135 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2qwb\" (UniqueName: \"kubernetes.io/projected/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575-kube-api-access-g2qwb\") pod \"auto-csr-approver-29562890-xwcc2\" (UID: \"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575\") " pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257163 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-proxy-ca-bundles\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257204 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-serving-cert\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257229 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jbqhm\" (UniqueName: \"kubernetes.io/projected/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-kube-api-access-jbqhm\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257252 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9nktg\" (UniqueName: \"kubernetes.io/projected/da230aff-847c-4f62-8f1c-06f126239d1b-kube-api-access-9nktg\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257289 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da230aff-847c-4f62-8f1c-06f126239d1b-serving-cert\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257328 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-config\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257370 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-config\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257392 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-tmp\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257447 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c75701-692d-4bb5-a0d5-2020c652ab82-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257463 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8r5j7\" (UniqueName: \"kubernetes.io/projected/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-kube-api-access-8r5j7\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257477 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4s9r9\" (UniqueName: \"kubernetes.io/projected/85c75701-692d-4bb5-a0d5-2020c652ab82-kube-api-access-4s9r9\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.257488 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.258024 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-tmp\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.259555 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da230aff-847c-4f62-8f1c-06f126239d1b-tmp\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.259778 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-config\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.259898 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-client-ca\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.260072 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-config\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.260529 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-proxy-ca-bundles\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.264860 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" event={"ID":"85c75701-692d-4bb5-a0d5-2020c652ab82","Type":"ContainerDied","Data":"797be210166a85a6cd44dfe58626d57eaf7f4c11cee1440db3f51fb2cf8696fa"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.264990 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-lz6ks" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.266471 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-client-ca\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.269982 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-serving-cert\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.270772 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerStarted","Data":"ba41597b2de44ff6eb5676959ed286eebe14a53151a6aa7da42775d0d2290eef"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.271494 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da230aff-847c-4f62-8f1c-06f126239d1b-serving-cert\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.278389 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" event={"ID":"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141","Type":"ContainerStarted","Data":"48677fd6b89217942625a8b8d9a775d95f8ffe38dea479476c4b4a9bd2296bb5"} Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.280378 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbqhm\" (UniqueName: \"kubernetes.io/projected/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-kube-api-access-jbqhm\") pod \"controller-manager-6fddb65cd-rtxpl\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.297867 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nktg\" (UniqueName: \"kubernetes.io/projected/da230aff-847c-4f62-8f1c-06f126239d1b-kube-api-access-9nktg\") pod \"route-controller-manager-94465b4f4-92w8s\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.300896 5110 scope.go:117] "RemoveContainer" containerID="54a40f6fcadf19fd5733bfcd79d1e67f1f317709bf00d13f8d23ad9bd8bc8df8" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.302049 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.307160 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64d44f6ddf-8jwk2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.358629 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g2qwb\" (UniqueName: \"kubernetes.io/projected/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575-kube-api-access-g2qwb\") pod \"auto-csr-approver-29562890-xwcc2\" (UID: \"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575\") " pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.391937 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-lz6ks"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.393609 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2qwb\" (UniqueName: \"kubernetes.io/projected/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575-kube-api-access-g2qwb\") pod \"auto-csr-approver-29562890-xwcc2\" (UID: \"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575\") " pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.398947 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-lz6ks"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.402521 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.429495 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.433443 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.442735 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.528440 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.753008 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c75701-692d-4bb5-a0d5-2020c652ab82" path="/var/lib/kubelet/pods/85c75701-692d-4bb5-a0d5-2020c652ab82/volumes" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.754593 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" path="/var/lib/kubelet/pods/fc6f0b62-c129-4c6d-a228-af2e48b6a0ce/volumes" Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.837568 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fddb65cd-rtxpl"] Mar 17 18:50:00 crc kubenswrapper[5110]: W0317 18:50:00.850281 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b4add05_6452_4ffa_afd9_56ea9a6b14e9.slice/crio-420eeccccb39058255a2e5148a7e34b87e2882a720a1e9621fecacfa352da790 WatchSource:0}: Error finding container 420eeccccb39058255a2e5148a7e34b87e2882a720a1e9621fecacfa352da790: Status 404 returned error can't find the container with id 420eeccccb39058255a2e5148a7e34b87e2882a720a1e9621fecacfa352da790 Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.957398 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562890-xwcc2"] Mar 17 18:50:00 crc kubenswrapper[5110]: I0317 18:50:00.960324 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s"] Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.022948 5110 patch_prober.go:28] interesting pod/route-controller-manager-776cdc94d6-m9q7t container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": context deadline exceeded" start-of-body= Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.023017 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-m9q7t" podUID="fc6f0b62-c129-4c6d-a228-af2e48b6a0ce" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": context deadline exceeded" Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.284635 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" event={"ID":"9b4add05-6452-4ffa-afd9-56ea9a6b14e9","Type":"ContainerStarted","Data":"420eeccccb39058255a2e5148a7e34b87e2882a720a1e9621fecacfa352da790"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.286563 5110 generic.go:358] "Generic (PLEG): container finished" podID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerID="381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.286633 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx699" event={"ID":"c15d66c3-0baf-44b5-aa64-ff26d681badd","Type":"ContainerDied","Data":"381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.288637 5110 generic.go:358] "Generic (PLEG): container finished" podID="9fd9df75-65d5-476e-9e82-da35f1116996" containerID="0ecc64867c73b49e8d7fd6b12ce965282df07aaa3c03a503afb080d528c217b2" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.288722 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh5rg" event={"ID":"9fd9df75-65d5-476e-9e82-da35f1116996","Type":"ContainerDied","Data":"0ecc64867c73b49e8d7fd6b12ce965282df07aaa3c03a503afb080d528c217b2"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.290291 5110 generic.go:358] "Generic (PLEG): container finished" podID="ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8" containerID="e9c7f776cb8ba4e8e614d8f19d6ac6fa36ee19bb50687b20f8fd8fa4bfcfed85" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.290332 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8","Type":"ContainerDied","Data":"e9c7f776cb8ba4e8e614d8f19d6ac6fa36ee19bb50687b20f8fd8fa4bfcfed85"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.294419 5110 generic.go:358] "Generic (PLEG): container finished" podID="5f27c690-db4e-404f-88d9-60f93687021d" containerID="9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.294499 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerDied","Data":"9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.301959 5110 generic.go:358] "Generic (PLEG): container finished" podID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerID="2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.302090 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerDied","Data":"2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.307978 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerStarted","Data":"fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.314779 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" event={"ID":"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575","Type":"ContainerStarted","Data":"1e3ade87a7d20fc8f6ba448d87e3532785841e101e505c9ceff41f3f95bdf4ea"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.316523 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" event={"ID":"da230aff-847c-4f62-8f1c-06f126239d1b","Type":"ContainerStarted","Data":"f687b24c2f904b97568328c659609b9e9d96cc0a94a6ba7b8576e04184bdce72"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.321982 5110 generic.go:358] "Generic (PLEG): container finished" podID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerID="6eed216aebd0ab34f7b96032e984f7c4b7620698696d48a20bd480014b415183" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.322101 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czltc" event={"ID":"a0b63bb8-f8c1-4591-affd-65e1d6cb3393","Type":"ContainerDied","Data":"6eed216aebd0ab34f7b96032e984f7c4b7620698696d48a20bd480014b415183"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.327611 5110 generic.go:358] "Generic (PLEG): container finished" podID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerID="b7664d5b2c2744a1bd9fe29f53fdbd2c44c70ae044074eec0151f5fedd1b1447" exitCode=0 Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.327732 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldd2" event={"ID":"72bdf6fb-83a5-4dcb-996c-0bee47903667","Type":"ContainerDied","Data":"b7664d5b2c2744a1bd9fe29f53fdbd2c44c70ae044074eec0151f5fedd1b1447"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.331097 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m5g58" event={"ID":"baee355c-9055-42f1-87c4-3dc7a08b68f6","Type":"ContainerStarted","Data":"571d794f5abc8b1e286accb8680530070d59b1e9be05ab0e350a7d79f06b9f4d"} Mar 17 18:50:01 crc kubenswrapper[5110]: I0317 18:50:01.453716 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-m5g58" podStartSLOduration=124.453697551 podStartE2EDuration="2m4.453697551s" podCreationTimestamp="2026-03-17 18:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:01.452090387 +0000 UTC m=+145.484701909" watchObservedRunningTime="2026-03-17 18:50:01.453697551 +0000 UTC m=+145.486309073" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.341552 5110 generic.go:358] "Generic (PLEG): container finished" podID="2b3686e6-f0ef-4197-867d-67617aff597e" containerID="ba41597b2de44ff6eb5676959ed286eebe14a53151a6aa7da42775d0d2290eef" exitCode=0 Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.341630 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerDied","Data":"ba41597b2de44ff6eb5676959ed286eebe14a53151a6aa7da42775d0d2290eef"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.344686 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" event={"ID":"9b4add05-6452-4ffa-afd9-56ea9a6b14e9","Type":"ContainerStarted","Data":"883d8990d7f6c8d7e1e67f1059b58fa18dc4df254e20e4c2e9338e5f22a8b4be"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.344918 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.347687 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx699" event={"ID":"c15d66c3-0baf-44b5-aa64-ff26d681badd","Type":"ContainerStarted","Data":"f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.350160 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh5rg" event={"ID":"9fd9df75-65d5-476e-9e82-da35f1116996","Type":"ContainerStarted","Data":"2808e465d7090bf5dc3dabbb865ad0a0bc592a6ff26bb912931ee02f7f4af634"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.351946 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerStarted","Data":"be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.352554 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.353792 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerStarted","Data":"b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.357237 5110 generic.go:358] "Generic (PLEG): container finished" podID="66594344-3d62-414a-9e89-3ab843c5239a" containerID="fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c" exitCode=0 Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.357330 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerDied","Data":"fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.359223 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" event={"ID":"da230aff-847c-4f62-8f1c-06f126239d1b","Type":"ContainerStarted","Data":"246a73edfb4e0146aac05c4353b23023bfd0e8a610cf05c6478d85fed4cd3ea7"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.362085 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czltc" event={"ID":"a0b63bb8-f8c1-4591-affd-65e1d6cb3393","Type":"ContainerStarted","Data":"052c29ca4712916b5105c422bb78da12e80470450fd5d29aeda39dd4bef5d3e6"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.368740 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldd2" event={"ID":"72bdf6fb-83a5-4dcb-996c-0bee47903667","Type":"ContainerStarted","Data":"5ddd0af732769f3a01a5f513e450a9a6ed4a6589e2c0cb2a9bfb40f8d8b14e56"} Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.391269 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" podStartSLOduration=11.39124888 podStartE2EDuration="11.39124888s" podCreationTimestamp="2026-03-17 18:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:02.389126701 +0000 UTC m=+146.421738243" watchObservedRunningTime="2026-03-17 18:50:02.39124888 +0000 UTC m=+146.423860402" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.394591 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.456105 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.506149 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bh5rg" podStartSLOduration=4.516883804 podStartE2EDuration="24.506128316s" podCreationTimestamp="2026-03-17 18:49:38 +0000 UTC" firstStartedPulling="2026-03-17 18:49:39.910207258 +0000 UTC m=+123.942818780" lastFinishedPulling="2026-03-17 18:49:59.89945177 +0000 UTC m=+143.932063292" observedRunningTime="2026-03-17 18:50:02.486472899 +0000 UTC m=+146.519084421" watchObservedRunningTime="2026-03-17 18:50:02.506128316 +0000 UTC m=+146.538739828" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.542221 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zw8tx" podStartSLOduration=4.605018871 podStartE2EDuration="23.542202459s" podCreationTimestamp="2026-03-17 18:49:39 +0000 UTC" firstStartedPulling="2026-03-17 18:49:40.959127864 +0000 UTC m=+124.991739386" lastFinishedPulling="2026-03-17 18:49:59.896311452 +0000 UTC m=+143.928922974" observedRunningTime="2026-03-17 18:50:02.539991287 +0000 UTC m=+146.572602809" watchObservedRunningTime="2026-03-17 18:50:02.542202459 +0000 UTC m=+146.574813991" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.542347 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-czltc" podStartSLOduration=5.448805085 podStartE2EDuration="26.542342483s" podCreationTimestamp="2026-03-17 18:49:36 +0000 UTC" firstStartedPulling="2026-03-17 18:49:38.802760584 +0000 UTC m=+122.835372106" lastFinishedPulling="2026-03-17 18:49:59.896297982 +0000 UTC m=+143.928909504" observedRunningTime="2026-03-17 18:50:02.521475902 +0000 UTC m=+146.554087424" watchObservedRunningTime="2026-03-17 18:50:02.542342483 +0000 UTC m=+146.574954005" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.587822 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rx699" podStartSLOduration=5.592898137 podStartE2EDuration="25.587804337s" podCreationTimestamp="2026-03-17 18:49:37 +0000 UTC" firstStartedPulling="2026-03-17 18:49:39.901814204 +0000 UTC m=+123.934425726" lastFinishedPulling="2026-03-17 18:49:59.896720404 +0000 UTC m=+143.929331926" observedRunningTime="2026-03-17 18:50:02.566079803 +0000 UTC m=+146.598691335" watchObservedRunningTime="2026-03-17 18:50:02.587804337 +0000 UTC m=+146.620415859" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.615990 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" podStartSLOduration=11.615972321 podStartE2EDuration="11.615972321s" podCreationTimestamp="2026-03-17 18:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:02.590894263 +0000 UTC m=+146.623505805" watchObservedRunningTime="2026-03-17 18:50:02.615972321 +0000 UTC m=+146.648583843" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.638770 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nldd2" podStartSLOduration=5.624025889 podStartE2EDuration="26.638755025s" podCreationTimestamp="2026-03-17 18:49:36 +0000 UTC" firstStartedPulling="2026-03-17 18:49:38.810825488 +0000 UTC m=+122.843437010" lastFinishedPulling="2026-03-17 18:49:59.825554624 +0000 UTC m=+143.858166146" observedRunningTime="2026-03-17 18:50:02.617521414 +0000 UTC m=+146.650132936" watchObservedRunningTime="2026-03-17 18:50:02.638755025 +0000 UTC m=+146.671366627" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.639278 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4vpt2" podStartSLOduration=4.67656801 podStartE2EDuration="25.639272489s" podCreationTimestamp="2026-03-17 18:49:37 +0000 UTC" firstStartedPulling="2026-03-17 18:49:38.862601638 +0000 UTC m=+122.895213160" lastFinishedPulling="2026-03-17 18:49:59.825306117 +0000 UTC m=+143.857917639" observedRunningTime="2026-03-17 18:50:02.638877188 +0000 UTC m=+146.671488720" watchObservedRunningTime="2026-03-17 18:50:02.639272489 +0000 UTC m=+146.671884011" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.733299 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.856159 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kube-api-access\") pod \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.856273 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kubelet-dir\") pod \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\" (UID: \"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8\") " Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.856454 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8" (UID: "ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.865341 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8" (UID: "ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.957609 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:02 crc kubenswrapper[5110]: I0317 18:50:02.957647 5110 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.376488 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerStarted","Data":"a9fabb5949652a9641e847a094d265787a384b4616857c9f88fefff64f32599f"} Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.378906 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8","Type":"ContainerDied","Data":"802665a648d6c611942fe87141f3610275a3adedc744e489d00463c47e992942"} Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.378933 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="802665a648d6c611942fe87141f3610275a3adedc744e489d00463c47e992942" Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.379004 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.388117 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerStarted","Data":"3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9"} Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.425048 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-955md" podStartSLOduration=5.571477789 podStartE2EDuration="23.425028355s" podCreationTimestamp="2026-03-17 18:49:40 +0000 UTC" firstStartedPulling="2026-03-17 18:49:42.028840909 +0000 UTC m=+126.061452431" lastFinishedPulling="2026-03-17 18:49:59.882391475 +0000 UTC m=+143.915002997" observedRunningTime="2026-03-17 18:50:03.407989471 +0000 UTC m=+147.440601003" watchObservedRunningTime="2026-03-17 18:50:03.425028355 +0000 UTC m=+147.457639877" Mar 17 18:50:03 crc kubenswrapper[5110]: I0317 18:50:03.427109 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jctgs" podStartSLOduration=6.448086777 podStartE2EDuration="24.427103253s" podCreationTimestamp="2026-03-17 18:49:39 +0000 UTC" firstStartedPulling="2026-03-17 18:49:42.044214117 +0000 UTC m=+126.076825639" lastFinishedPulling="2026-03-17 18:50:00.023230603 +0000 UTC m=+144.055842115" observedRunningTime="2026-03-17 18:50:03.423218675 +0000 UTC m=+147.455830197" watchObservedRunningTime="2026-03-17 18:50:03.427103253 +0000 UTC m=+147.459714775" Mar 17 18:50:04 crc kubenswrapper[5110]: I0317 18:50:04.110973 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:50:04 crc kubenswrapper[5110]: E0317 18:50:04.299823 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:50:04 crc kubenswrapper[5110]: E0317 18:50:04.309861 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:50:04 crc kubenswrapper[5110]: E0317 18:50:04.315797 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 17 18:50:04 crc kubenswrapper[5110]: E0317 18:50:04.315864 5110 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.268433 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.269363 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.272584 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.272657 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.288490 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-jqrnn_10e07288-52f4-407c-a8eb-709214e358f5/kube-multus-additional-cni-plugins/0.log" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.288610 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.412799 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-jqrnn_10e07288-52f4-407c-a8eb-709214e358f5/kube-multus-additional-cni-plugins/0.log" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.413040 5110 generic.go:358] "Generic (PLEG): container finished" podID="10e07288-52f4-407c-a8eb-709214e358f5" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" exitCode=137 Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.413136 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" event={"ID":"10e07288-52f4-407c-a8eb-709214e358f5","Type":"ContainerDied","Data":"6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a"} Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.413201 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.413218 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jqrnn" event={"ID":"10e07288-52f4-407c-a8eb-709214e358f5","Type":"ContainerDied","Data":"c33605948ab99822eeeef3dad82ac7fe459817fb201ec525b04509b9dd961700"} Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.413248 5110 scope.go:117] "RemoveContainer" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424190 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jm87\" (UniqueName: \"kubernetes.io/projected/10e07288-52f4-407c-a8eb-709214e358f5-kube-api-access-2jm87\") pod \"10e07288-52f4-407c-a8eb-709214e358f5\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424289 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10e07288-52f4-407c-a8eb-709214e358f5-cni-sysctl-allowlist\") pod \"10e07288-52f4-407c-a8eb-709214e358f5\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424358 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10e07288-52f4-407c-a8eb-709214e358f5-tuning-conf-dir\") pod \"10e07288-52f4-407c-a8eb-709214e358f5\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424410 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/10e07288-52f4-407c-a8eb-709214e358f5-ready\") pod \"10e07288-52f4-407c-a8eb-709214e358f5\" (UID: \"10e07288-52f4-407c-a8eb-709214e358f5\") " Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424533 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10e07288-52f4-407c-a8eb-709214e358f5-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "10e07288-52f4-407c-a8eb-709214e358f5" (UID: "10e07288-52f4-407c-a8eb-709214e358f5"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424878 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10e07288-52f4-407c-a8eb-709214e358f5-ready" (OuterVolumeSpecName: "ready") pod "10e07288-52f4-407c-a8eb-709214e358f5" (UID: "10e07288-52f4-407c-a8eb-709214e358f5"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.424900 5110 reconciler_common.go:299] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10e07288-52f4-407c-a8eb-709214e358f5-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.425175 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e07288-52f4-407c-a8eb-709214e358f5-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "10e07288-52f4-407c-a8eb-709214e358f5" (UID: "10e07288-52f4-407c-a8eb-709214e358f5"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.430809 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10e07288-52f4-407c-a8eb-709214e358f5-kube-api-access-2jm87" (OuterVolumeSpecName: "kube-api-access-2jm87") pod "10e07288-52f4-407c-a8eb-709214e358f5" (UID: "10e07288-52f4-407c-a8eb-709214e358f5"). InnerVolumeSpecName "kube-api-access-2jm87". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.439035 5110 scope.go:117] "RemoveContainer" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" Mar 17 18:50:07 crc kubenswrapper[5110]: E0317 18:50:07.439817 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a\": container with ID starting with 6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a not found: ID does not exist" containerID="6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.439866 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a"} err="failed to get container status \"6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a\": rpc error: code = NotFound desc = could not find container \"6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a\": container with ID starting with 6df9692d136fb8346ec18622815620f84dc30bda9c3121e023da0c5a13362f2a not found: ID does not exist" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.526610 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2jm87\" (UniqueName: \"kubernetes.io/projected/10e07288-52f4-407c-a8eb-709214e358f5-kube-api-access-2jm87\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.526657 5110 reconciler_common.go:299] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10e07288-52f4-407c-a8eb-709214e358f5-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.526672 5110 reconciler_common.go:299] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/10e07288-52f4-407c-a8eb-709214e358f5-ready\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.734730 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.734795 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.743948 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jqrnn"] Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.752586 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jqrnn"] Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.795910 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-gg6pg" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.894467 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.894581 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.894658 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.897031 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.897085 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.951956 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.965738 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:50:07 crc kubenswrapper[5110]: I0317 18:50:07.990425 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:50:08 crc kubenswrapper[5110]: I0317 18:50:08.461963 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:50:08 crc kubenswrapper[5110]: I0317 18:50:08.471140 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:50:08 crc kubenswrapper[5110]: I0317 18:50:08.752834 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10e07288-52f4-407c-a8eb-709214e358f5" path="/var/lib/kubelet/pods/10e07288-52f4-407c-a8eb-709214e358f5/volumes" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.071897 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.072311 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.130735 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.395243 5110 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-9wk4n" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.402228 5110 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-9wk4n" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.428766 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.428965 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.430797 5110 generic.go:358] "Generic (PLEG): container finished" podID="44f42740-7b5b-4e4f-9e9e-c4c83b6c1575" containerID="f4b4b2b33f22225cd1c4788078e50e75b65f3a5923ea97ad909adbe4ec3e98e8" exitCode=0 Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.430999 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" event={"ID":"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575","Type":"ContainerDied","Data":"f4b4b2b33f22225cd1c4788078e50e75b65f3a5923ea97ad909adbe4ec3e98e8"} Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.483311 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.485763 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:50:09 crc kubenswrapper[5110]: I0317 18:50:09.665147 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4vpt2"] Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.239706 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.239779 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.287466 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.403862 5110 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2026-04-16 18:45:09 +0000 UTC" deadline="2026-04-07 23:31:00.22105709 +0000 UTC" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.404136 5110 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="508h40m49.816927157s" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.440863 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4vpt2" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="registry-server" containerID="cri-o://be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a" gracePeriod=2 Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.485846 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.491112 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.610127 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.611128 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.658495 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.806416 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.874547 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.971253 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2qwb\" (UniqueName: \"kubernetes.io/projected/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575-kube-api-access-g2qwb\") pod \"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575\" (UID: \"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575\") " Mar 17 18:50:10 crc kubenswrapper[5110]: I0317 18:50:10.977717 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575-kube-api-access-g2qwb" (OuterVolumeSpecName: "kube-api-access-g2qwb") pod "44f42740-7b5b-4e4f-9e9e-c4c83b6c1575" (UID: "44f42740-7b5b-4e4f-9e9e-c4c83b6c1575"). InnerVolumeSpecName "kube-api-access-g2qwb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.058988 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6fddb65cd-rtxpl"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.059424 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" podUID="9b4add05-6452-4ffa-afd9-56ea9a6b14e9" containerName="controller-manager" containerID="cri-o://883d8990d7f6c8d7e1e67f1059b58fa18dc4df254e20e4c2e9338e5f22a8b4be" gracePeriod=30 Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.062666 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.063261 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" podUID="da230aff-847c-4f62-8f1c-06f126239d1b" containerName="route-controller-manager" containerID="cri-o://246a73edfb4e0146aac05c4353b23023bfd0e8a610cf05c6478d85fed4cd3ea7" gracePeriod=30 Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.072211 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-775bm\" (UniqueName: \"kubernetes.io/projected/5f27c690-db4e-404f-88d9-60f93687021d-kube-api-access-775bm\") pod \"5f27c690-db4e-404f-88d9-60f93687021d\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.072359 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-catalog-content\") pod \"5f27c690-db4e-404f-88d9-60f93687021d\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.072408 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-utilities\") pod \"5f27c690-db4e-404f-88d9-60f93687021d\" (UID: \"5f27c690-db4e-404f-88d9-60f93687021d\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.072683 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-g2qwb\" (UniqueName: \"kubernetes.io/projected/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575-kube-api-access-g2qwb\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.073480 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-utilities" (OuterVolumeSpecName: "utilities") pod "5f27c690-db4e-404f-88d9-60f93687021d" (UID: "5f27c690-db4e-404f-88d9-60f93687021d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.078096 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f27c690-db4e-404f-88d9-60f93687021d-kube-api-access-775bm" (OuterVolumeSpecName: "kube-api-access-775bm") pod "5f27c690-db4e-404f-88d9-60f93687021d" (UID: "5f27c690-db4e-404f-88d9-60f93687021d"). InnerVolumeSpecName "kube-api-access-775bm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.106200 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f27c690-db4e-404f-88d9-60f93687021d" (UID: "5f27c690-db4e-404f-88d9-60f93687021d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.173674 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-775bm\" (UniqueName: \"kubernetes.io/projected/5f27c690-db4e-404f-88d9-60f93687021d-kube-api-access-775bm\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.173705 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.173713 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f27c690-db4e-404f-88d9-60f93687021d-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.451157 5110 generic.go:358] "Generic (PLEG): container finished" podID="da230aff-847c-4f62-8f1c-06f126239d1b" containerID="246a73edfb4e0146aac05c4353b23023bfd0e8a610cf05c6478d85fed4cd3ea7" exitCode=0 Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.451202 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" event={"ID":"da230aff-847c-4f62-8f1c-06f126239d1b","Type":"ContainerDied","Data":"246a73edfb4e0146aac05c4353b23023bfd0e8a610cf05c6478d85fed4cd3ea7"} Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.452700 5110 generic.go:358] "Generic (PLEG): container finished" podID="9b4add05-6452-4ffa-afd9-56ea9a6b14e9" containerID="883d8990d7f6c8d7e1e67f1059b58fa18dc4df254e20e4c2e9338e5f22a8b4be" exitCode=0 Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.452723 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" event={"ID":"9b4add05-6452-4ffa-afd9-56ea9a6b14e9","Type":"ContainerDied","Data":"883d8990d7f6c8d7e1e67f1059b58fa18dc4df254e20e4c2e9338e5f22a8b4be"} Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.454656 5110 generic.go:358] "Generic (PLEG): container finished" podID="5f27c690-db4e-404f-88d9-60f93687021d" containerID="be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a" exitCode=0 Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.454716 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerDied","Data":"be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a"} Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.454741 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4vpt2" event={"ID":"5f27c690-db4e-404f-88d9-60f93687021d","Type":"ContainerDied","Data":"3a136d18084e4722f55612ff0d1cfed91420e9b4216935a0c98d27e67fddae7a"} Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.454748 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4vpt2" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.454758 5110 scope.go:117] "RemoveContainer" containerID="be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.458431 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" event={"ID":"44f42740-7b5b-4e4f-9e9e-c4c83b6c1575","Type":"ContainerDied","Data":"1e3ade87a7d20fc8f6ba448d87e3532785841e101e505c9ceff41f3f95bdf4ea"} Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.458471 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e3ade87a7d20fc8f6ba448d87e3532785841e101e505c9ceff41f3f95bdf4ea" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.458655 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562890-xwcc2" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.466648 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rx699"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.467046 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rx699" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="registry-server" containerID="cri-o://f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67" gracePeriod=2 Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.482646 5110 scope.go:117] "RemoveContainer" containerID="9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.502405 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.506249 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4vpt2"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.518519 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4vpt2"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.521011 5110 scope.go:117] "RemoveContainer" containerID="ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.537134 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538111 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538639 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="da230aff-847c-4f62-8f1c-06f126239d1b" containerName="route-controller-manager" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538656 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="da230aff-847c-4f62-8f1c-06f126239d1b" containerName="route-controller-manager" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538667 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8" containerName="pruner" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538673 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8" containerName="pruner" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538684 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="44f42740-7b5b-4e4f-9e9e-c4c83b6c1575" containerName="oc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538690 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f42740-7b5b-4e4f-9e9e-c4c83b6c1575" containerName="oc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538713 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="registry-server" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538718 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="registry-server" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538742 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="extract-utilities" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538748 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="extract-utilities" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538757 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="extract-content" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538763 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="extract-content" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538771 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538777 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538856 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="da230aff-847c-4f62-8f1c-06f126239d1b" containerName="route-controller-manager" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538867 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="5f27c690-db4e-404f-88d9-60f93687021d" containerName="registry-server" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538877 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="ccaa6003-d5d6-4eea-b2ac-7178f4cc7ec8" containerName="pruner" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538887 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="44f42740-7b5b-4e4f-9e9e-c4c83b6c1575" containerName="oc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.538894 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="10e07288-52f4-407c-a8eb-709214e358f5" containerName="kube-multus-additional-cni-plugins" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.553456 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.553604 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.565123 5110 scope.go:117] "RemoveContainer" containerID="be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a" Mar 17 18:50:11 crc kubenswrapper[5110]: E0317 18:50:11.568520 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a\": container with ID starting with be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a not found: ID does not exist" containerID="be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.568572 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a"} err="failed to get container status \"be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a\": rpc error: code = NotFound desc = could not find container \"be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a\": container with ID starting with be89249d471079d0b4319a9e35a0b4f2f28e671272fba04181becfa6c1ec406a not found: ID does not exist" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.568600 5110 scope.go:117] "RemoveContainer" containerID="9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef" Mar 17 18:50:11 crc kubenswrapper[5110]: E0317 18:50:11.569235 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef\": container with ID starting with 9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef not found: ID does not exist" containerID="9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.569290 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef"} err="failed to get container status \"9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef\": rpc error: code = NotFound desc = could not find container \"9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef\": container with ID starting with 9a452a70486291ec3120af21ad20d7696db1c45baea47a5d696420650b8d36ef not found: ID does not exist" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.569324 5110 scope.go:117] "RemoveContainer" containerID="ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d" Mar 17 18:50:11 crc kubenswrapper[5110]: E0317 18:50:11.569871 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d\": container with ID starting with ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d not found: ID does not exist" containerID="ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.569906 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d"} err="failed to get container status \"ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d\": rpc error: code = NotFound desc = could not find container \"ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d\": container with ID starting with ad63b21fdbeeedfc8ab629dbecca2efe9606ab99c0ea2b8025067f20aab57f5d not found: ID does not exist" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587493 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-client-ca\") pod \"da230aff-847c-4f62-8f1c-06f126239d1b\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587548 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da230aff-847c-4f62-8f1c-06f126239d1b-serving-cert\") pod \"da230aff-847c-4f62-8f1c-06f126239d1b\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587587 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-config\") pod \"da230aff-847c-4f62-8f1c-06f126239d1b\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587622 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da230aff-847c-4f62-8f1c-06f126239d1b-tmp\") pod \"da230aff-847c-4f62-8f1c-06f126239d1b\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587683 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nktg\" (UniqueName: \"kubernetes.io/projected/da230aff-847c-4f62-8f1c-06f126239d1b-kube-api-access-9nktg\") pod \"da230aff-847c-4f62-8f1c-06f126239d1b\" (UID: \"da230aff-847c-4f62-8f1c-06f126239d1b\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587803 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ea90c130-d78e-46c3-ae4e-ab5a54281576-tmp\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587845 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-config\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587881 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-client-ca\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587919 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9lfb\" (UniqueName: \"kubernetes.io/projected/ea90c130-d78e-46c3-ae4e-ab5a54281576-kube-api-access-q9lfb\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.587945 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea90c130-d78e-46c3-ae4e-ab5a54281576-serving-cert\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.588709 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-client-ca" (OuterVolumeSpecName: "client-ca") pod "da230aff-847c-4f62-8f1c-06f126239d1b" (UID: "da230aff-847c-4f62-8f1c-06f126239d1b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.592005 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da230aff-847c-4f62-8f1c-06f126239d1b-tmp" (OuterVolumeSpecName: "tmp") pod "da230aff-847c-4f62-8f1c-06f126239d1b" (UID: "da230aff-847c-4f62-8f1c-06f126239d1b"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.592086 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-config" (OuterVolumeSpecName: "config") pod "da230aff-847c-4f62-8f1c-06f126239d1b" (UID: "da230aff-847c-4f62-8f1c-06f126239d1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.595244 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da230aff-847c-4f62-8f1c-06f126239d1b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "da230aff-847c-4f62-8f1c-06f126239d1b" (UID: "da230aff-847c-4f62-8f1c-06f126239d1b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.595423 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da230aff-847c-4f62-8f1c-06f126239d1b-kube-api-access-9nktg" (OuterVolumeSpecName: "kube-api-access-9nktg") pod "da230aff-847c-4f62-8f1c-06f126239d1b" (UID: "da230aff-847c-4f62-8f1c-06f126239d1b"). InnerVolumeSpecName "kube-api-access-9nktg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.643598 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.667659 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56b68954dc-4898r"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.668230 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9b4add05-6452-4ffa-afd9-56ea9a6b14e9" containerName="controller-manager" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.668243 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4add05-6452-4ffa-afd9-56ea9a6b14e9" containerName="controller-manager" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.668361 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="9b4add05-6452-4ffa-afd9-56ea9a6b14e9" containerName="controller-manager" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.679435 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.687785 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56b68954dc-4898r"] Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.688708 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbqhm\" (UniqueName: \"kubernetes.io/projected/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-kube-api-access-jbqhm\") pod \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.688785 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-config\") pod \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.688840 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-proxy-ca-bundles\") pod \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.688893 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-serving-cert\") pod \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.688924 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-tmp\") pod \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.688965 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-client-ca\") pod \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\" (UID: \"9b4add05-6452-4ffa-afd9-56ea9a6b14e9\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.689041 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-config\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.689723 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-tmp" (OuterVolumeSpecName: "tmp") pod "9b4add05-6452-4ffa-afd9-56ea9a6b14e9" (UID: "9b4add05-6452-4ffa-afd9-56ea9a6b14e9"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.689872 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-config\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.689932 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-client-ca\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690146 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b4add05-6452-4ffa-afd9-56ea9a6b14e9" (UID: "9b4add05-6452-4ffa-afd9-56ea9a6b14e9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690252 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-client-ca\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690322 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q9lfb\" (UniqueName: \"kubernetes.io/projected/ea90c130-d78e-46c3-ae4e-ab5a54281576-kube-api-access-q9lfb\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690346 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ccbb9488-7532-4984-bc4b-4ef3077244ff-tmp\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690403 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea90c130-d78e-46c3-ae4e-ab5a54281576-serving-cert\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690506 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-proxy-ca-bundles\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690590 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sngb2\" (UniqueName: \"kubernetes.io/projected/ccbb9488-7532-4984-bc4b-4ef3077244ff-kube-api-access-sngb2\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690627 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ea90c130-d78e-46c3-ae4e-ab5a54281576-tmp\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690671 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbb9488-7532-4984-bc4b-4ef3077244ff-serving-cert\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690744 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690757 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da230aff-847c-4f62-8f1c-06f126239d1b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690770 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da230aff-847c-4f62-8f1c-06f126239d1b-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690781 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/da230aff-847c-4f62-8f1c-06f126239d1b-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690794 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9nktg\" (UniqueName: \"kubernetes.io/projected/da230aff-847c-4f62-8f1c-06f126239d1b-kube-api-access-9nktg\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690809 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.690820 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.691132 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-config" (OuterVolumeSpecName: "config") pod "9b4add05-6452-4ffa-afd9-56ea9a6b14e9" (UID: "9b4add05-6452-4ffa-afd9-56ea9a6b14e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.692186 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ea90c130-d78e-46c3-ae4e-ab5a54281576-tmp\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.692922 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-client-ca\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.693118 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9b4add05-6452-4ffa-afd9-56ea9a6b14e9" (UID: "9b4add05-6452-4ffa-afd9-56ea9a6b14e9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.693262 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-config\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.694170 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-kube-api-access-jbqhm" (OuterVolumeSpecName: "kube-api-access-jbqhm") pod "9b4add05-6452-4ffa-afd9-56ea9a6b14e9" (UID: "9b4add05-6452-4ffa-afd9-56ea9a6b14e9"). InnerVolumeSpecName "kube-api-access-jbqhm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.697019 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea90c130-d78e-46c3-ae4e-ab5a54281576-serving-cert\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.699132 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b4add05-6452-4ffa-afd9-56ea9a6b14e9" (UID: "9b4add05-6452-4ffa-afd9-56ea9a6b14e9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.707337 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9lfb\" (UniqueName: \"kubernetes.io/projected/ea90c130-d78e-46c3-ae4e-ab5a54281576-kube-api-access-q9lfb\") pod \"route-controller-manager-df9dfdc67-rqtqc\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.744282 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.791958 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-utilities\") pod \"c15d66c3-0baf-44b5-aa64-ff26d681badd\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792024 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-catalog-content\") pod \"c15d66c3-0baf-44b5-aa64-ff26d681badd\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792120 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gjt2\" (UniqueName: \"kubernetes.io/projected/c15d66c3-0baf-44b5-aa64-ff26d681badd-kube-api-access-9gjt2\") pod \"c15d66c3-0baf-44b5-aa64-ff26d681badd\" (UID: \"c15d66c3-0baf-44b5-aa64-ff26d681badd\") " Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792250 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbb9488-7532-4984-bc4b-4ef3077244ff-serving-cert\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792311 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-config\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792490 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-client-ca\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792578 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ccbb9488-7532-4984-bc4b-4ef3077244ff-tmp\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792688 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-proxy-ca-bundles\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792741 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sngb2\" (UniqueName: \"kubernetes.io/projected/ccbb9488-7532-4984-bc4b-4ef3077244ff-kube-api-access-sngb2\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792824 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-jbqhm\" (UniqueName: \"kubernetes.io/projected/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-kube-api-access-jbqhm\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792841 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792851 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.792860 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b4add05-6452-4ffa-afd9-56ea9a6b14e9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.793440 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ccbb9488-7532-4984-bc4b-4ef3077244ff-tmp\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.793567 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-utilities" (OuterVolumeSpecName: "utilities") pod "c15d66c3-0baf-44b5-aa64-ff26d681badd" (UID: "c15d66c3-0baf-44b5-aa64-ff26d681badd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.793603 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-client-ca\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.794330 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-proxy-ca-bundles\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.794893 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-config\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.796208 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15d66c3-0baf-44b5-aa64-ff26d681badd-kube-api-access-9gjt2" (OuterVolumeSpecName: "kube-api-access-9gjt2") pod "c15d66c3-0baf-44b5-aa64-ff26d681badd" (UID: "c15d66c3-0baf-44b5-aa64-ff26d681badd"). InnerVolumeSpecName "kube-api-access-9gjt2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.796985 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbb9488-7532-4984-bc4b-4ef3077244ff-serving-cert\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.809925 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sngb2\" (UniqueName: \"kubernetes.io/projected/ccbb9488-7532-4984-bc4b-4ef3077244ff-kube-api-access-sngb2\") pod \"controller-manager-56b68954dc-4898r\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.858896 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c15d66c3-0baf-44b5-aa64-ff26d681badd" (UID: "c15d66c3-0baf-44b5-aa64-ff26d681badd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.893880 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.893902 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15d66c3-0baf-44b5-aa64-ff26d681badd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.893914 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9gjt2\" (UniqueName: \"kubernetes.io/projected/c15d66c3-0baf-44b5-aa64-ff26d681badd-kube-api-access-9gjt2\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:11 crc kubenswrapper[5110]: I0317 18:50:11.940731 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.015753 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.067321 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zw8tx"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.250560 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56b68954dc-4898r"] Mar 17 18:50:12 crc kubenswrapper[5110]: W0317 18:50:12.257123 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccbb9488_7532_4984_bc4b_4ef3077244ff.slice/crio-c1633d7836cf4205f032da07ed9a6c0df3a07c207a3c3972a0f507c5b749fd29 WatchSource:0}: Error finding container c1633d7836cf4205f032da07ed9a6c0df3a07c207a3c3972a0f507c5b749fd29: Status 404 returned error can't find the container with id c1633d7836cf4205f032da07ed9a6c0df3a07c207a3c3972a0f507c5b749fd29 Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.369468 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc"] Mar 17 18:50:12 crc kubenswrapper[5110]: W0317 18:50:12.375698 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea90c130_d78e_46c3_ae4e_ab5a54281576.slice/crio-5dccc8da8a93ada9253667727182555b581ff3b839bfd7b4711730f4adbe5e26 WatchSource:0}: Error finding container 5dccc8da8a93ada9253667727182555b581ff3b839bfd7b4711730f4adbe5e26: Status 404 returned error can't find the container with id 5dccc8da8a93ada9253667727182555b581ff3b839bfd7b4711730f4adbe5e26 Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.465433 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" event={"ID":"9b4add05-6452-4ffa-afd9-56ea9a6b14e9","Type":"ContainerDied","Data":"420eeccccb39058255a2e5148a7e34b87e2882a720a1e9621fecacfa352da790"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.465962 5110 scope.go:117] "RemoveContainer" containerID="883d8990d7f6c8d7e1e67f1059b58fa18dc4df254e20e4c2e9338e5f22a8b4be" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.466039 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fddb65cd-rtxpl" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.476515 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx699" event={"ID":"c15d66c3-0baf-44b5-aa64-ff26d681badd","Type":"ContainerDied","Data":"f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.476463 5110 generic.go:358] "Generic (PLEG): container finished" podID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerID="f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67" exitCode=0 Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.476710 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx699" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.476816 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx699" event={"ID":"c15d66c3-0baf-44b5-aa64-ff26d681badd","Type":"ContainerDied","Data":"8ec46c9c115dcdd6b68e0b2e35eaa2fa87336d0be3a444fcd9106f07252b51c3"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.485816 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" event={"ID":"da230aff-847c-4f62-8f1c-06f126239d1b","Type":"ContainerDied","Data":"f687b24c2f904b97568328c659609b9e9d96cc0a94a6ba7b8576e04184bdce72"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.485830 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.488940 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" event={"ID":"ea90c130-d78e-46c3-ae4e-ab5a54281576","Type":"ContainerStarted","Data":"5dccc8da8a93ada9253667727182555b581ff3b839bfd7b4711730f4adbe5e26"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.490000 5110 scope.go:117] "RemoveContainer" containerID="f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.493623 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" event={"ID":"ccbb9488-7532-4984-bc4b-4ef3077244ff","Type":"ContainerStarted","Data":"507525e9f537096e6190ed96adb851d784f4021ca2dbced49b516b03617f4043"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.493661 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" event={"ID":"ccbb9488-7532-4984-bc4b-4ef3077244ff","Type":"ContainerStarted","Data":"c1633d7836cf4205f032da07ed9a6c0df3a07c207a3c3972a0f507c5b749fd29"} Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.493877 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zw8tx" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="registry-server" containerID="cri-o://b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f" gracePeriod=2 Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.494136 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.509628 5110 scope.go:117] "RemoveContainer" containerID="381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.526124 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" podStartSLOduration=1.526104457 podStartE2EDuration="1.526104457s" podCreationTimestamp="2026-03-17 18:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:12.51437617 +0000 UTC m=+156.546987692" watchObservedRunningTime="2026-03-17 18:50:12.526104457 +0000 UTC m=+156.558715979" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.527526 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.534443 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-94465b4f4-92w8s"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.536882 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6fddb65cd-rtxpl"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.539772 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6fddb65cd-rtxpl"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.548851 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rx699"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.548896 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rx699"] Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.601206 5110 scope.go:117] "RemoveContainer" containerID="502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.626846 5110 scope.go:117] "RemoveContainer" containerID="f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67" Mar 17 18:50:12 crc kubenswrapper[5110]: E0317 18:50:12.627326 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67\": container with ID starting with f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67 not found: ID does not exist" containerID="f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.627376 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67"} err="failed to get container status \"f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67\": rpc error: code = NotFound desc = could not find container \"f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67\": container with ID starting with f52ee075784725847fed0d5b94d2a6a5d9f1e1361cf0dd9087510f4c7fd96b67 not found: ID does not exist" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.627405 5110 scope.go:117] "RemoveContainer" containerID="381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9" Mar 17 18:50:12 crc kubenswrapper[5110]: E0317 18:50:12.627747 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9\": container with ID starting with 381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9 not found: ID does not exist" containerID="381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.627816 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9"} err="failed to get container status \"381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9\": rpc error: code = NotFound desc = could not find container \"381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9\": container with ID starting with 381b6f6c00e7396311a61194842aa4ef94f669624f331f1a93af768aa3feccf9 not found: ID does not exist" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.627844 5110 scope.go:117] "RemoveContainer" containerID="502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4" Mar 17 18:50:12 crc kubenswrapper[5110]: E0317 18:50:12.628163 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4\": container with ID starting with 502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4 not found: ID does not exist" containerID="502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.628183 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4"} err="failed to get container status \"502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4\": rpc error: code = NotFound desc = could not find container \"502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4\": container with ID starting with 502c41fe5861f1784af729d276ce0bab8078bf5172c415e59e935742e4292ca4 not found: ID does not exist" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.628196 5110 scope.go:117] "RemoveContainer" containerID="246a73edfb4e0146aac05c4353b23023bfd0e8a610cf05c6478d85fed4cd3ea7" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.760946 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f27c690-db4e-404f-88d9-60f93687021d" path="/var/lib/kubelet/pods/5f27c690-db4e-404f-88d9-60f93687021d/volumes" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.761924 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b4add05-6452-4ffa-afd9-56ea9a6b14e9" path="/var/lib/kubelet/pods/9b4add05-6452-4ffa-afd9-56ea9a6b14e9/volumes" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.762372 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" path="/var/lib/kubelet/pods/c15d66c3-0baf-44b5-aa64-ff26d681badd/volumes" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.763509 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da230aff-847c-4f62-8f1c-06f126239d1b" path="/var/lib/kubelet/pods/da230aff-847c-4f62-8f1c-06f126239d1b/volumes" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.816407 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.923537 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-utilities\") pod \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.923630 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-catalog-content\") pod \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.923664 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kshp\" (UniqueName: \"kubernetes.io/projected/ba6a6e05-30d4-4581-95ac-3ca772fad52e-kube-api-access-2kshp\") pod \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\" (UID: \"ba6a6e05-30d4-4581-95ac-3ca772fad52e\") " Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.924434 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-utilities" (OuterVolumeSpecName: "utilities") pod "ba6a6e05-30d4-4581-95ac-3ca772fad52e" (UID: "ba6a6e05-30d4-4581-95ac-3ca772fad52e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.929703 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6a6e05-30d4-4581-95ac-3ca772fad52e-kube-api-access-2kshp" (OuterVolumeSpecName: "kube-api-access-2kshp") pod "ba6a6e05-30d4-4581-95ac-3ca772fad52e" (UID: "ba6a6e05-30d4-4581-95ac-3ca772fad52e"). InnerVolumeSpecName "kube-api-access-2kshp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.937965 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba6a6e05-30d4-4581-95ac-3ca772fad52e" (UID: "ba6a6e05-30d4-4581-95ac-3ca772fad52e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:12 crc kubenswrapper[5110]: I0317 18:50:12.959350 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.025112 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2kshp\" (UniqueName: \"kubernetes.io/projected/ba6a6e05-30d4-4581-95ac-3ca772fad52e-kube-api-access-2kshp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.025138 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.025147 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6a6e05-30d4-4581-95ac-3ca772fad52e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.510818 5110 generic.go:358] "Generic (PLEG): container finished" podID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerID="b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f" exitCode=0 Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.511128 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerDied","Data":"b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f"} Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.511197 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zw8tx" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.511317 5110 scope.go:117] "RemoveContainer" containerID="b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.511301 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zw8tx" event={"ID":"ba6a6e05-30d4-4581-95ac-3ca772fad52e","Type":"ContainerDied","Data":"77dce939691c0e8e2393fc89a96d0251f64e27b0c697d82e1f8506e049773e20"} Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.516324 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" event={"ID":"ea90c130-d78e-46c3-ae4e-ab5a54281576","Type":"ContainerStarted","Data":"7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f"} Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.517551 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.522443 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.539394 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" podStartSLOduration=2.539378192 podStartE2EDuration="2.539378192s" podCreationTimestamp="2026-03-17 18:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:13.535142674 +0000 UTC m=+157.567754216" watchObservedRunningTime="2026-03-17 18:50:13.539378192 +0000 UTC m=+157.571989714" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.548727 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zw8tx"] Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.550577 5110 scope.go:117] "RemoveContainer" containerID="2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.552487 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zw8tx"] Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.574981 5110 scope.go:117] "RemoveContainer" containerID="14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.600254 5110 scope.go:117] "RemoveContainer" containerID="b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f" Mar 17 18:50:13 crc kubenswrapper[5110]: E0317 18:50:13.600709 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f\": container with ID starting with b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f not found: ID does not exist" containerID="b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.600822 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f"} err="failed to get container status \"b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f\": rpc error: code = NotFound desc = could not find container \"b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f\": container with ID starting with b34442f47035307dab36c1210c10e2dd5f9155e4f9b0a22030f3917b8f9d446f not found: ID does not exist" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.600913 5110 scope.go:117] "RemoveContainer" containerID="2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04" Mar 17 18:50:13 crc kubenswrapper[5110]: E0317 18:50:13.601465 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04\": container with ID starting with 2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04 not found: ID does not exist" containerID="2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.601507 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04"} err="failed to get container status \"2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04\": rpc error: code = NotFound desc = could not find container \"2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04\": container with ID starting with 2e74bec91005e3eb5c46307ad99c533b317faa819b326d0cb050703c78b1ab04 not found: ID does not exist" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.601531 5110 scope.go:117] "RemoveContainer" containerID="14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816" Mar 17 18:50:13 crc kubenswrapper[5110]: E0317 18:50:13.601922 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816\": container with ID starting with 14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816 not found: ID does not exist" containerID="14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816" Mar 17 18:50:13 crc kubenswrapper[5110]: I0317 18:50:13.602003 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816"} err="failed to get container status \"14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816\": rpc error: code = NotFound desc = could not find container \"14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816\": container with ID starting with 14835c913f58b34ac2be898b4779b79234835e06c3107ad95f0b34d81d00c816 not found: ID does not exist" Mar 17 18:50:14 crc kubenswrapper[5110]: I0317 18:50:14.469028 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-955md"] Mar 17 18:50:14 crc kubenswrapper[5110]: I0317 18:50:14.469374 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-955md" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="registry-server" containerID="cri-o://a9fabb5949652a9641e847a094d265787a384b4616857c9f88fefff64f32599f" gracePeriod=2 Mar 17 18:50:14 crc kubenswrapper[5110]: I0317 18:50:14.753144 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" path="/var/lib/kubelet/pods/ba6a6e05-30d4-4581-95ac-3ca772fad52e/volumes" Mar 17 18:50:15 crc kubenswrapper[5110]: I0317 18:50:15.532899 5110 generic.go:358] "Generic (PLEG): container finished" podID="2b3686e6-f0ef-4197-867d-67617aff597e" containerID="a9fabb5949652a9641e847a094d265787a384b4616857c9f88fefff64f32599f" exitCode=0 Mar 17 18:50:15 crc kubenswrapper[5110]: I0317 18:50:15.533781 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerDied","Data":"a9fabb5949652a9641e847a094d265787a384b4616857c9f88fefff64f32599f"} Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.180159 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.451179 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.539845 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-955md" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.539858 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-955md" event={"ID":"2b3686e6-f0ef-4197-867d-67617aff597e","Type":"ContainerDied","Data":"af911c1b82538ea5188357c043eb789df4f0389cb9f41dea59cf8decdea10e4e"} Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.539943 5110 scope.go:117] "RemoveContainer" containerID="a9fabb5949652a9641e847a094d265787a384b4616857c9f88fefff64f32599f" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.558738 5110 scope.go:117] "RemoveContainer" containerID="ba41597b2de44ff6eb5676959ed286eebe14a53151a6aa7da42775d0d2290eef" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.569609 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxm47\" (UniqueName: \"kubernetes.io/projected/2b3686e6-f0ef-4197-867d-67617aff597e-kube-api-access-rxm47\") pod \"2b3686e6-f0ef-4197-867d-67617aff597e\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.569657 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-utilities\") pod \"2b3686e6-f0ef-4197-867d-67617aff597e\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.569689 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-catalog-content\") pod \"2b3686e6-f0ef-4197-867d-67617aff597e\" (UID: \"2b3686e6-f0ef-4197-867d-67617aff597e\") " Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.572751 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-utilities" (OuterVolumeSpecName: "utilities") pod "2b3686e6-f0ef-4197-867d-67617aff597e" (UID: "2b3686e6-f0ef-4197-867d-67617aff597e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.574824 5110 scope.go:117] "RemoveContainer" containerID="24bb633ad12b8741d94fa92085e649159aebcc5d4eafc288a70c7dbf83add4c6" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.582184 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3686e6-f0ef-4197-867d-67617aff597e-kube-api-access-rxm47" (OuterVolumeSpecName: "kube-api-access-rxm47") pod "2b3686e6-f0ef-4197-867d-67617aff597e" (UID: "2b3686e6-f0ef-4197-867d-67617aff597e"). InnerVolumeSpecName "kube-api-access-rxm47". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.671354 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rxm47\" (UniqueName: \"kubernetes.io/projected/2b3686e6-f0ef-4197-867d-67617aff597e-kube-api-access-rxm47\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.671396 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.683131 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b3686e6-f0ef-4197-867d-67617aff597e" (UID: "2b3686e6-f0ef-4197-867d-67617aff597e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.780813 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3686e6-f0ef-4197-867d-67617aff597e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.853939 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-955md"] Mar 17 18:50:16 crc kubenswrapper[5110]: I0317 18:50:16.860041 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-955md"] Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840422 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840923 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="extract-content" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840935 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="extract-content" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840943 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840949 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840962 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="extract-utilities" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840967 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="extract-utilities" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840974 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="extract-content" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840979 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="extract-content" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.840992 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841000 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841014 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="extract-utilities" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841019 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="extract-utilities" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841026 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841031 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841038 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="extract-content" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841044 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="extract-content" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841077 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="extract-utilities" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841083 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="extract-utilities" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841160 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="c15d66c3-0baf-44b5-aa64-ff26d681badd" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841170 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.841177 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="ba6a6e05-30d4-4581-95ac-3ca772fad52e" containerName="registry-server" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.846337 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.854871 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.855979 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.856204 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.893815 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e13257d-6532-47cf-be25-502b6bdfddad-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.893881 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e13257d-6532-47cf-be25-502b6bdfddad-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.995319 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e13257d-6532-47cf-be25-502b6bdfddad-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.995360 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e13257d-6532-47cf-be25-502b6bdfddad-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:17 crc kubenswrapper[5110]: I0317 18:50:17.995456 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e13257d-6532-47cf-be25-502b6bdfddad-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:18 crc kubenswrapper[5110]: I0317 18:50:18.017507 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e13257d-6532-47cf-be25-502b6bdfddad-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:18 crc kubenswrapper[5110]: I0317 18:50:18.169115 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:18 crc kubenswrapper[5110]: I0317 18:50:18.562638 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Mar 17 18:50:18 crc kubenswrapper[5110]: I0317 18:50:18.752872 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b3686e6-f0ef-4197-867d-67617aff597e" path="/var/lib/kubelet/pods/2b3686e6-f0ef-4197-867d-67617aff597e/volumes" Mar 17 18:50:19 crc kubenswrapper[5110]: I0317 18:50:19.572912 5110 generic.go:358] "Generic (PLEG): container finished" podID="5e13257d-6532-47cf-be25-502b6bdfddad" containerID="6da09869080e729953224819b8b897924b029b95136d7fae9041554f2184cd01" exitCode=0 Mar 17 18:50:19 crc kubenswrapper[5110]: I0317 18:50:19.572976 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"5e13257d-6532-47cf-be25-502b6bdfddad","Type":"ContainerDied","Data":"6da09869080e729953224819b8b897924b029b95136d7fae9041554f2184cd01"} Mar 17 18:50:19 crc kubenswrapper[5110]: I0317 18:50:19.573236 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"5e13257d-6532-47cf-be25-502b6bdfddad","Type":"ContainerStarted","Data":"083570147b319a8e0cfe18e5b688d8ce38ff5a5c6bd5a9733e7fd2a321c1f804"} Mar 17 18:50:20 crc kubenswrapper[5110]: I0317 18:50:20.875292 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:20 crc kubenswrapper[5110]: I0317 18:50:20.930819 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e13257d-6532-47cf-be25-502b6bdfddad-kube-api-access\") pod \"5e13257d-6532-47cf-be25-502b6bdfddad\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " Mar 17 18:50:20 crc kubenswrapper[5110]: I0317 18:50:20.930951 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e13257d-6532-47cf-be25-502b6bdfddad-kubelet-dir\") pod \"5e13257d-6532-47cf-be25-502b6bdfddad\" (UID: \"5e13257d-6532-47cf-be25-502b6bdfddad\") " Mar 17 18:50:20 crc kubenswrapper[5110]: I0317 18:50:20.931195 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e13257d-6532-47cf-be25-502b6bdfddad-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5e13257d-6532-47cf-be25-502b6bdfddad" (UID: "5e13257d-6532-47cf-be25-502b6bdfddad"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:50:20 crc kubenswrapper[5110]: I0317 18:50:20.937311 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e13257d-6532-47cf-be25-502b6bdfddad-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5e13257d-6532-47cf-be25-502b6bdfddad" (UID: "5e13257d-6532-47cf-be25-502b6bdfddad"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:21 crc kubenswrapper[5110]: I0317 18:50:21.032346 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e13257d-6532-47cf-be25-502b6bdfddad-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:21 crc kubenswrapper[5110]: I0317 18:50:21.032374 5110 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e13257d-6532-47cf-be25-502b6bdfddad-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:21 crc kubenswrapper[5110]: I0317 18:50:21.587283 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 17 18:50:21 crc kubenswrapper[5110]: I0317 18:50:21.587336 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"5e13257d-6532-47cf-be25-502b6bdfddad","Type":"ContainerDied","Data":"083570147b319a8e0cfe18e5b688d8ce38ff5a5c6bd5a9733e7fd2a321c1f804"} Mar 17 18:50:21 crc kubenswrapper[5110]: I0317 18:50:21.587896 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="083570147b319a8e0cfe18e5b688d8ce38ff5a5c6bd5a9733e7fd2a321c1f804" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.239532 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.240033 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5e13257d-6532-47cf-be25-502b6bdfddad" containerName="pruner" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.240045 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e13257d-6532-47cf-be25-502b6bdfddad" containerName="pruner" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.240205 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="5e13257d-6532-47cf-be25-502b6bdfddad" containerName="pruner" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.246192 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.249887 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.250014 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.258962 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.288366 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kube-api-access\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.288425 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-var-lock\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.288475 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kubelet-dir\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.389597 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kube-api-access\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.389659 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-var-lock\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.389713 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kubelet-dir\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.389791 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kubelet-dir\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.389831 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-var-lock\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.409877 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kube-api-access\") pod \"installer-12-crc\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.561656 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:50:25 crc kubenswrapper[5110]: I0317 18:50:25.965938 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Mar 17 18:50:25 crc kubenswrapper[5110]: W0317 18:50:25.970774 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7ecce33d_988f_4c1d_b3ee_37c9e949a3df.slice/crio-b18d32500bcdc3cc6a9dc085ee1037348503ea6e8df5671d11469b3eccb5dcb5 WatchSource:0}: Error finding container b18d32500bcdc3cc6a9dc085ee1037348503ea6e8df5671d11469b3eccb5dcb5: Status 404 returned error can't find the container with id b18d32500bcdc3cc6a9dc085ee1037348503ea6e8df5671d11469b3eccb5dcb5 Mar 17 18:50:26 crc kubenswrapper[5110]: I0317 18:50:26.618634 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7ecce33d-988f-4c1d-b3ee-37c9e949a3df","Type":"ContainerStarted","Data":"7af855336ae1f74151b8d1db4c8cb8eaed795808011b6bbde668a0b81c5535a5"} Mar 17 18:50:26 crc kubenswrapper[5110]: I0317 18:50:26.618684 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7ecce33d-988f-4c1d-b3ee-37c9e949a3df","Type":"ContainerStarted","Data":"b18d32500bcdc3cc6a9dc085ee1037348503ea6e8df5671d11469b3eccb5dcb5"} Mar 17 18:50:26 crc kubenswrapper[5110]: I0317 18:50:26.634224 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-12-crc" podStartSLOduration=1.63420776 podStartE2EDuration="1.63420776s" podCreationTimestamp="2026-03-17 18:50:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:26.632314778 +0000 UTC m=+170.664926300" watchObservedRunningTime="2026-03-17 18:50:26.63420776 +0000 UTC m=+170.666819282" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.067465 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56b68954dc-4898r"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.069331 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" podUID="ccbb9488-7532-4984-bc4b-4ef3077244ff" containerName="controller-manager" containerID="cri-o://507525e9f537096e6190ed96adb851d784f4021ca2dbced49b516b03617f4043" gracePeriod=30 Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.090728 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.091191 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" podUID="ea90c130-d78e-46c3-ae4e-ab5a54281576" containerName="route-controller-manager" containerID="cri-o://7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f" gracePeriod=30 Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.590228 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.620729 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.621455 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ea90c130-d78e-46c3-ae4e-ab5a54281576" containerName="route-controller-manager" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.621488 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea90c130-d78e-46c3-ae4e-ab5a54281576" containerName="route-controller-manager" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.621598 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="ea90c130-d78e-46c3-ae4e-ab5a54281576" containerName="route-controller-manager" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.627953 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.628200 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.654774 5110 generic.go:358] "Generic (PLEG): container finished" podID="ea90c130-d78e-46c3-ae4e-ab5a54281576" containerID="7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f" exitCode=0 Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.654906 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" event={"ID":"ea90c130-d78e-46c3-ae4e-ab5a54281576","Type":"ContainerDied","Data":"7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f"} Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.654940 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" event={"ID":"ea90c130-d78e-46c3-ae4e-ab5a54281576","Type":"ContainerDied","Data":"5dccc8da8a93ada9253667727182555b581ff3b839bfd7b4711730f4adbe5e26"} Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.654959 5110 scope.go:117] "RemoveContainer" containerID="7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.655136 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.658395 5110 generic.go:358] "Generic (PLEG): container finished" podID="ccbb9488-7532-4984-bc4b-4ef3077244ff" containerID="507525e9f537096e6190ed96adb851d784f4021ca2dbced49b516b03617f4043" exitCode=0 Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.658498 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" event={"ID":"ccbb9488-7532-4984-bc4b-4ef3077244ff","Type":"ContainerDied","Data":"507525e9f537096e6190ed96adb851d784f4021ca2dbced49b516b03617f4043"} Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.672795 5110 scope.go:117] "RemoveContainer" containerID="7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f" Mar 17 18:50:31 crc kubenswrapper[5110]: E0317 18:50:31.673227 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f\": container with ID starting with 7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f not found: ID does not exist" containerID="7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.673286 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f"} err="failed to get container status \"7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f\": rpc error: code = NotFound desc = could not find container \"7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f\": container with ID starting with 7bc93f32cc547ff7865b4618966e7c02beacdc4053f7866e16afa3caf228bf5f not found: ID does not exist" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.716850 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ea90c130-d78e-46c3-ae4e-ab5a54281576-tmp\") pod \"ea90c130-d78e-46c3-ae4e-ab5a54281576\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717307 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea90c130-d78e-46c3-ae4e-ab5a54281576-serving-cert\") pod \"ea90c130-d78e-46c3-ae4e-ab5a54281576\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717422 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-client-ca\") pod \"ea90c130-d78e-46c3-ae4e-ab5a54281576\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717429 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea90c130-d78e-46c3-ae4e-ab5a54281576-tmp" (OuterVolumeSpecName: "tmp") pod "ea90c130-d78e-46c3-ae4e-ab5a54281576" (UID: "ea90c130-d78e-46c3-ae4e-ab5a54281576"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717509 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-config\") pod \"ea90c130-d78e-46c3-ae4e-ab5a54281576\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717563 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9lfb\" (UniqueName: \"kubernetes.io/projected/ea90c130-d78e-46c3-ae4e-ab5a54281576-kube-api-access-q9lfb\") pod \"ea90c130-d78e-46c3-ae4e-ab5a54281576\" (UID: \"ea90c130-d78e-46c3-ae4e-ab5a54281576\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717719 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8423c766-29a0-455c-a843-554ed1b96c71-tmp\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717783 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlxlq\" (UniqueName: \"kubernetes.io/projected/8423c766-29a0-455c-a843-554ed1b96c71-kube-api-access-mlxlq\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717850 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-client-ca\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717906 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-client-ca" (OuterVolumeSpecName: "client-ca") pod "ea90c130-d78e-46c3-ae4e-ab5a54281576" (UID: "ea90c130-d78e-46c3-ae4e-ab5a54281576"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.717973 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-config\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.718008 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8423c766-29a0-455c-a843-554ed1b96c71-serving-cert\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.718176 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ea90c130-d78e-46c3-ae4e-ab5a54281576-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.718200 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.718253 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-config" (OuterVolumeSpecName: "config") pod "ea90c130-d78e-46c3-ae4e-ab5a54281576" (UID: "ea90c130-d78e-46c3-ae4e-ab5a54281576"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.722746 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea90c130-d78e-46c3-ae4e-ab5a54281576-kube-api-access-q9lfb" (OuterVolumeSpecName: "kube-api-access-q9lfb") pod "ea90c130-d78e-46c3-ae4e-ab5a54281576" (UID: "ea90c130-d78e-46c3-ae4e-ab5a54281576"). InnerVolumeSpecName "kube-api-access-q9lfb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.723325 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea90c130-d78e-46c3-ae4e-ab5a54281576-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ea90c130-d78e-46c3-ae4e-ab5a54281576" (UID: "ea90c130-d78e-46c3-ae4e-ab5a54281576"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.727377 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.765925 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77478b5f5-lvvct"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.767289 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ccbb9488-7532-4984-bc4b-4ef3077244ff" containerName="controller-manager" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.767327 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbb9488-7532-4984-bc4b-4ef3077244ff" containerName="controller-manager" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.767703 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="ccbb9488-7532-4984-bc4b-4ef3077244ff" containerName="controller-manager" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.790709 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77478b5f5-lvvct"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.790928 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819592 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbb9488-7532-4984-bc4b-4ef3077244ff-serving-cert\") pod \"ccbb9488-7532-4984-bc4b-4ef3077244ff\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819646 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-config\") pod \"ccbb9488-7532-4984-bc4b-4ef3077244ff\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819676 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sngb2\" (UniqueName: \"kubernetes.io/projected/ccbb9488-7532-4984-bc4b-4ef3077244ff-kube-api-access-sngb2\") pod \"ccbb9488-7532-4984-bc4b-4ef3077244ff\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819757 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ccbb9488-7532-4984-bc4b-4ef3077244ff-tmp\") pod \"ccbb9488-7532-4984-bc4b-4ef3077244ff\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819772 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-proxy-ca-bundles\") pod \"ccbb9488-7532-4984-bc4b-4ef3077244ff\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819834 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-client-ca\") pod \"ccbb9488-7532-4984-bc4b-4ef3077244ff\" (UID: \"ccbb9488-7532-4984-bc4b-4ef3077244ff\") " Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.819987 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8423c766-29a0-455c-a843-554ed1b96c71-tmp\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820028 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mlxlq\" (UniqueName: \"kubernetes.io/projected/8423c766-29a0-455c-a843-554ed1b96c71-kube-api-access-mlxlq\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820049 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-client-ca\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820094 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-config\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820111 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8423c766-29a0-455c-a843-554ed1b96c71-serving-cert\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820181 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea90c130-d78e-46c3-ae4e-ab5a54281576-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820192 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea90c130-d78e-46c3-ae4e-ab5a54281576-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.820200 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q9lfb\" (UniqueName: \"kubernetes.io/projected/ea90c130-d78e-46c3-ae4e-ab5a54281576-kube-api-access-q9lfb\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.821357 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ccbb9488-7532-4984-bc4b-4ef3077244ff" (UID: "ccbb9488-7532-4984-bc4b-4ef3077244ff"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.821386 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-config" (OuterVolumeSpecName: "config") pod "ccbb9488-7532-4984-bc4b-4ef3077244ff" (UID: "ccbb9488-7532-4984-bc4b-4ef3077244ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.821683 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbb9488-7532-4984-bc4b-4ef3077244ff-tmp" (OuterVolumeSpecName: "tmp") pod "ccbb9488-7532-4984-bc4b-4ef3077244ff" (UID: "ccbb9488-7532-4984-bc4b-4ef3077244ff"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.822252 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-client-ca" (OuterVolumeSpecName: "client-ca") pod "ccbb9488-7532-4984-bc4b-4ef3077244ff" (UID: "ccbb9488-7532-4984-bc4b-4ef3077244ff"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.822537 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8423c766-29a0-455c-a843-554ed1b96c71-tmp\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.822889 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-client-ca\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.824323 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8423c766-29a0-455c-a843-554ed1b96c71-serving-cert\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.825113 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbb9488-7532-4984-bc4b-4ef3077244ff-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ccbb9488-7532-4984-bc4b-4ef3077244ff" (UID: "ccbb9488-7532-4984-bc4b-4ef3077244ff"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.827113 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-config\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.827353 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccbb9488-7532-4984-bc4b-4ef3077244ff-kube-api-access-sngb2" (OuterVolumeSpecName: "kube-api-access-sngb2") pod "ccbb9488-7532-4984-bc4b-4ef3077244ff" (UID: "ccbb9488-7532-4984-bc4b-4ef3077244ff"). InnerVolumeSpecName "kube-api-access-sngb2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.838384 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlxlq\" (UniqueName: \"kubernetes.io/projected/8423c766-29a0-455c-a843-554ed1b96c71-kube-api-access-mlxlq\") pod \"route-controller-manager-6dd7b4bdbb-zbdrk\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.921669 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-client-ca\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.921801 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/853aa33a-fe6b-4972-a94b-077ab373a5b2-tmp\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.921847 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-config\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922148 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853aa33a-fe6b-4972-a94b-077ab373a5b2-serving-cert\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922209 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-proxy-ca-bundles\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922237 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzqqk\" (UniqueName: \"kubernetes.io/projected/853aa33a-fe6b-4972-a94b-077ab373a5b2-kube-api-access-pzqqk\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922395 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922420 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbb9488-7532-4984-bc4b-4ef3077244ff-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922432 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922444 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sngb2\" (UniqueName: \"kubernetes.io/projected/ccbb9488-7532-4984-bc4b-4ef3077244ff-kube-api-access-sngb2\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922457 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ccbb9488-7532-4984-bc4b-4ef3077244ff-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.922469 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccbb9488-7532-4984-bc4b-4ef3077244ff-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.945578 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.985189 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc"] Mar 17 18:50:31 crc kubenswrapper[5110]: I0317 18:50:31.988248 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-df9dfdc67-rqtqc"] Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.023805 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853aa33a-fe6b-4972-a94b-077ab373a5b2-serving-cert\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.023854 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-proxy-ca-bundles\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.023875 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pzqqk\" (UniqueName: \"kubernetes.io/projected/853aa33a-fe6b-4972-a94b-077ab373a5b2-kube-api-access-pzqqk\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.023908 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-client-ca\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.023939 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/853aa33a-fe6b-4972-a94b-077ab373a5b2-tmp\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.023962 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-config\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.025473 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-proxy-ca-bundles\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.026093 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/853aa33a-fe6b-4972-a94b-077ab373a5b2-tmp\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.026210 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-config\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.027743 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-client-ca\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.029538 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853aa33a-fe6b-4972-a94b-077ab373a5b2-serving-cert\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.049203 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzqqk\" (UniqueName: \"kubernetes.io/projected/853aa33a-fe6b-4972-a94b-077ab373a5b2-kube-api-access-pzqqk\") pod \"controller-manager-77478b5f5-lvvct\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.109291 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.343917 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk"] Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.505022 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77478b5f5-lvvct"] Mar 17 18:50:32 crc kubenswrapper[5110]: W0317 18:50:32.517480 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod853aa33a_fe6b_4972_a94b_077ab373a5b2.slice/crio-8b2f81d7c2393b2ecfc4d38cb9f759ccd30145e8b805f0bebd21fd26b57ae087 WatchSource:0}: Error finding container 8b2f81d7c2393b2ecfc4d38cb9f759ccd30145e8b805f0bebd21fd26b57ae087: Status 404 returned error can't find the container with id 8b2f81d7c2393b2ecfc4d38cb9f759ccd30145e8b805f0bebd21fd26b57ae087 Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.668466 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" event={"ID":"853aa33a-fe6b-4972-a94b-077ab373a5b2","Type":"ContainerStarted","Data":"8b2f81d7c2393b2ecfc4d38cb9f759ccd30145e8b805f0bebd21fd26b57ae087"} Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.672655 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" event={"ID":"ccbb9488-7532-4984-bc4b-4ef3077244ff","Type":"ContainerDied","Data":"c1633d7836cf4205f032da07ed9a6c0df3a07c207a3c3972a0f507c5b749fd29"} Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.672777 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b68954dc-4898r" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.672773 5110 scope.go:117] "RemoveContainer" containerID="507525e9f537096e6190ed96adb851d784f4021ca2dbced49b516b03617f4043" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.674665 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" event={"ID":"8423c766-29a0-455c-a843-554ed1b96c71","Type":"ContainerStarted","Data":"4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d"} Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.674733 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" event={"ID":"8423c766-29a0-455c-a843-554ed1b96c71","Type":"ContainerStarted","Data":"b57e47dd09f11aabfce766a5c91558ea0390733e2ac48ff0d11cc20f7dca351e"} Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.675913 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.702017 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" podStartSLOduration=1.7019958800000001 podStartE2EDuration="1.70199588s" podCreationTimestamp="2026-03-17 18:50:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:32.696739103 +0000 UTC m=+176.729350685" watchObservedRunningTime="2026-03-17 18:50:32.70199588 +0000 UTC m=+176.734607412" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.720615 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56b68954dc-4898r"] Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.724320 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-56b68954dc-4898r"] Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.753088 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccbb9488-7532-4984-bc4b-4ef3077244ff" path="/var/lib/kubelet/pods/ccbb9488-7532-4984-bc4b-4ef3077244ff/volumes" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.753953 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea90c130-d78e-46c3-ae4e-ab5a54281576" path="/var/lib/kubelet/pods/ea90c130-d78e-46c3-ae4e-ab5a54281576/volumes" Mar 17 18:50:32 crc kubenswrapper[5110]: I0317 18:50:32.956669 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:33 crc kubenswrapper[5110]: I0317 18:50:33.688666 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" event={"ID":"853aa33a-fe6b-4972-a94b-077ab373a5b2","Type":"ContainerStarted","Data":"ba4d082837844f2568db307913ca6c85a31c2f4959ebd5cc09dff086bc4b5f8f"} Mar 17 18:50:33 crc kubenswrapper[5110]: I0317 18:50:33.689456 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:33 crc kubenswrapper[5110]: I0317 18:50:33.698077 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:33 crc kubenswrapper[5110]: I0317 18:50:33.712784 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" podStartSLOduration=2.712766904 podStartE2EDuration="2.712766904s" podCreationTimestamp="2026-03-17 18:50:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:33.711444987 +0000 UTC m=+177.744056549" watchObservedRunningTime="2026-03-17 18:50:33.712766904 +0000 UTC m=+177.745378426" Mar 17 18:50:38 crc kubenswrapper[5110]: I0317 18:50:38.830558 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-zpddx"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.080436 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77478b5f5-lvvct"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.081279 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" podUID="853aa33a-fe6b-4972-a94b-077ab373a5b2" containerName="controller-manager" containerID="cri-o://ba4d082837844f2568db307913ca6c85a31c2f4959ebd5cc09dff086bc4b5f8f" gracePeriod=30 Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.101718 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.102532 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" podUID="8423c766-29a0-455c-a843-554ed1b96c71" containerName="route-controller-manager" containerID="cri-o://4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d" gracePeriod=30 Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.508607 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.532633 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.534124 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8423c766-29a0-455c-a843-554ed1b96c71" containerName="route-controller-manager" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.534148 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8423c766-29a0-455c-a843-554ed1b96c71" containerName="route-controller-manager" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.534286 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="8423c766-29a0-455c-a843-554ed1b96c71" containerName="route-controller-manager" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.543679 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.543838 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.601402 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8423c766-29a0-455c-a843-554ed1b96c71-serving-cert\") pod \"8423c766-29a0-455c-a843-554ed1b96c71\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.601500 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlxlq\" (UniqueName: \"kubernetes.io/projected/8423c766-29a0-455c-a843-554ed1b96c71-kube-api-access-mlxlq\") pod \"8423c766-29a0-455c-a843-554ed1b96c71\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.601525 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-client-ca\") pod \"8423c766-29a0-455c-a843-554ed1b96c71\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.601549 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8423c766-29a0-455c-a843-554ed1b96c71-tmp\") pod \"8423c766-29a0-455c-a843-554ed1b96c71\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.601616 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-config\") pod \"8423c766-29a0-455c-a843-554ed1b96c71\" (UID: \"8423c766-29a0-455c-a843-554ed1b96c71\") " Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.602426 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-config" (OuterVolumeSpecName: "config") pod "8423c766-29a0-455c-a843-554ed1b96c71" (UID: "8423c766-29a0-455c-a843-554ed1b96c71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.602652 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-client-ca" (OuterVolumeSpecName: "client-ca") pod "8423c766-29a0-455c-a843-554ed1b96c71" (UID: "8423c766-29a0-455c-a843-554ed1b96c71"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.602717 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8423c766-29a0-455c-a843-554ed1b96c71-tmp" (OuterVolumeSpecName: "tmp") pod "8423c766-29a0-455c-a843-554ed1b96c71" (UID: "8423c766-29a0-455c-a843-554ed1b96c71"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.607282 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8423c766-29a0-455c-a843-554ed1b96c71-kube-api-access-mlxlq" (OuterVolumeSpecName: "kube-api-access-mlxlq") pod "8423c766-29a0-455c-a843-554ed1b96c71" (UID: "8423c766-29a0-455c-a843-554ed1b96c71"). InnerVolumeSpecName "kube-api-access-mlxlq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.614879 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8423c766-29a0-455c-a843-554ed1b96c71-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8423c766-29a0-455c-a843-554ed1b96c71" (UID: "8423c766-29a0-455c-a843-554ed1b96c71"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703263 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54ef47e-80e3-4de6-855b-45657e28c18c-client-ca\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703309 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ef47e-80e3-4de6-855b-45657e28c18c-config\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703324 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmqjn\" (UniqueName: \"kubernetes.io/projected/a54ef47e-80e3-4de6-855b-45657e28c18c-kube-api-access-bmqjn\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703420 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54ef47e-80e3-4de6-855b-45657e28c18c-serving-cert\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703475 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a54ef47e-80e3-4de6-855b-45657e28c18c-tmp\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703600 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mlxlq\" (UniqueName: \"kubernetes.io/projected/8423c766-29a0-455c-a843-554ed1b96c71-kube-api-access-mlxlq\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703618 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703632 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8423c766-29a0-455c-a843-554ed1b96c71-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703640 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8423c766-29a0-455c-a843-554ed1b96c71-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.703649 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8423c766-29a0-455c-a843-554ed1b96c71-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.811756 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54ef47e-80e3-4de6-855b-45657e28c18c-client-ca\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.811819 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ef47e-80e3-4de6-855b-45657e28c18c-config\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.811848 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bmqjn\" (UniqueName: \"kubernetes.io/projected/a54ef47e-80e3-4de6-855b-45657e28c18c-kube-api-access-bmqjn\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.811902 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54ef47e-80e3-4de6-855b-45657e28c18c-serving-cert\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.811924 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a54ef47e-80e3-4de6-855b-45657e28c18c-tmp\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.812522 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a54ef47e-80e3-4de6-855b-45657e28c18c-tmp\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.814090 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ef47e-80e3-4de6-855b-45657e28c18c-config\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.814113 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54ef47e-80e3-4de6-855b-45657e28c18c-client-ca\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.818775 5110 generic.go:358] "Generic (PLEG): container finished" podID="8423c766-29a0-455c-a843-554ed1b96c71" containerID="4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d" exitCode=0 Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.818939 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.818960 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" event={"ID":"8423c766-29a0-455c-a843-554ed1b96c71","Type":"ContainerDied","Data":"4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d"} Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.819140 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk" event={"ID":"8423c766-29a0-455c-a843-554ed1b96c71","Type":"ContainerDied","Data":"b57e47dd09f11aabfce766a5c91558ea0390733e2ac48ff0d11cc20f7dca351e"} Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.819186 5110 scope.go:117] "RemoveContainer" containerID="4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.821194 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54ef47e-80e3-4de6-855b-45657e28c18c-serving-cert\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.821298 5110 generic.go:358] "Generic (PLEG): container finished" podID="853aa33a-fe6b-4972-a94b-077ab373a5b2" containerID="ba4d082837844f2568db307913ca6c85a31c2f4959ebd5cc09dff086bc4b5f8f" exitCode=0 Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.821314 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" event={"ID":"853aa33a-fe6b-4972-a94b-077ab373a5b2","Type":"ContainerDied","Data":"ba4d082837844f2568db307913ca6c85a31c2f4959ebd5cc09dff086bc4b5f8f"} Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.835522 5110 scope.go:117] "RemoveContainer" containerID="4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d" Mar 17 18:50:51 crc kubenswrapper[5110]: E0317 18:50:51.835916 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d\": container with ID starting with 4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d not found: ID does not exist" containerID="4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.835949 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d"} err="failed to get container status \"4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d\": rpc error: code = NotFound desc = could not find container \"4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d\": container with ID starting with 4772db01467ce9d4c7beb1e04051c2b8603f3727cefd9d8c5479e10ba90e412d not found: ID does not exist" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.850434 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmqjn\" (UniqueName: \"kubernetes.io/projected/a54ef47e-80e3-4de6-855b-45657e28c18c-kube-api-access-bmqjn\") pod \"route-controller-manager-6bf65448ff-xfwd6\" (UID: \"a54ef47e-80e3-4de6-855b-45657e28c18c\") " pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.853111 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.855633 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dd7b4bdbb-zbdrk"] Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.859249 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:51 crc kubenswrapper[5110]: I0317 18:50:51.983805 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.017253 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7797b97577-5df66"] Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.021815 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="853aa33a-fe6b-4972-a94b-077ab373a5b2" containerName="controller-manager" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.022161 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="853aa33a-fe6b-4972-a94b-077ab373a5b2" containerName="controller-manager" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.022673 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="853aa33a-fe6b-4972-a94b-077ab373a5b2" containerName="controller-manager" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.033281 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7797b97577-5df66"] Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.033486 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.078386 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6"] Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.121390 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-client-ca\") pod \"853aa33a-fe6b-4972-a94b-077ab373a5b2\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.121719 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853aa33a-fe6b-4972-a94b-077ab373a5b2-serving-cert\") pod \"853aa33a-fe6b-4972-a94b-077ab373a5b2\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.121769 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-config\") pod \"853aa33a-fe6b-4972-a94b-077ab373a5b2\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.121814 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/853aa33a-fe6b-4972-a94b-077ab373a5b2-tmp\") pod \"853aa33a-fe6b-4972-a94b-077ab373a5b2\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.121865 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-proxy-ca-bundles\") pod \"853aa33a-fe6b-4972-a94b-077ab373a5b2\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.121956 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzqqk\" (UniqueName: \"kubernetes.io/projected/853aa33a-fe6b-4972-a94b-077ab373a5b2-kube-api-access-pzqqk\") pod \"853aa33a-fe6b-4972-a94b-077ab373a5b2\" (UID: \"853aa33a-fe6b-4972-a94b-077ab373a5b2\") " Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122125 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-config\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122110 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/853aa33a-fe6b-4972-a94b-077ab373a5b2-tmp" (OuterVolumeSpecName: "tmp") pod "853aa33a-fe6b-4972-a94b-077ab373a5b2" (UID: "853aa33a-fe6b-4972-a94b-077ab373a5b2"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122292 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-client-ca\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122341 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24zfp\" (UniqueName: \"kubernetes.io/projected/47daee45-2558-42ed-a85c-a2a4f350b922-kube-api-access-24zfp\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122499 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47daee45-2558-42ed-a85c-a2a4f350b922-serving-cert\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122508 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "853aa33a-fe6b-4972-a94b-077ab373a5b2" (UID: "853aa33a-fe6b-4972-a94b-077ab373a5b2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122537 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-proxy-ca-bundles\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122756 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/47daee45-2558-42ed-a85c-a2a4f350b922-tmp\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122897 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/853aa33a-fe6b-4972-a94b-077ab373a5b2-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.122914 5110 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.123123 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-client-ca" (OuterVolumeSpecName: "client-ca") pod "853aa33a-fe6b-4972-a94b-077ab373a5b2" (UID: "853aa33a-fe6b-4972-a94b-077ab373a5b2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.123523 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-config" (OuterVolumeSpecName: "config") pod "853aa33a-fe6b-4972-a94b-077ab373a5b2" (UID: "853aa33a-fe6b-4972-a94b-077ab373a5b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.126012 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/853aa33a-fe6b-4972-a94b-077ab373a5b2-kube-api-access-pzqqk" (OuterVolumeSpecName: "kube-api-access-pzqqk") pod "853aa33a-fe6b-4972-a94b-077ab373a5b2" (UID: "853aa33a-fe6b-4972-a94b-077ab373a5b2"). InnerVolumeSpecName "kube-api-access-pzqqk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.126420 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/853aa33a-fe6b-4972-a94b-077ab373a5b2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "853aa33a-fe6b-4972-a94b-077ab373a5b2" (UID: "853aa33a-fe6b-4972-a94b-077ab373a5b2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224217 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47daee45-2558-42ed-a85c-a2a4f350b922-serving-cert\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224258 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-proxy-ca-bundles\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224279 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/47daee45-2558-42ed-a85c-a2a4f350b922-tmp\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224323 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-config\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224349 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-client-ca\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224374 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-24zfp\" (UniqueName: \"kubernetes.io/projected/47daee45-2558-42ed-a85c-a2a4f350b922-kube-api-access-24zfp\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224427 5110 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224437 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pzqqk\" (UniqueName: \"kubernetes.io/projected/853aa33a-fe6b-4972-a94b-077ab373a5b2-kube-api-access-pzqqk\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224446 5110 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853aa33a-fe6b-4972-a94b-077ab373a5b2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.224454 5110 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853aa33a-fe6b-4972-a94b-077ab373a5b2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.225101 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/47daee45-2558-42ed-a85c-a2a4f350b922-tmp\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.225619 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-client-ca\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.225721 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-proxy-ca-bundles\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.228972 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47daee45-2558-42ed-a85c-a2a4f350b922-config\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.232778 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47daee45-2558-42ed-a85c-a2a4f350b922-serving-cert\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.242458 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-24zfp\" (UniqueName: \"kubernetes.io/projected/47daee45-2558-42ed-a85c-a2a4f350b922-kube-api-access-24zfp\") pod \"controller-manager-7797b97577-5df66\" (UID: \"47daee45-2558-42ed-a85c-a2a4f350b922\") " pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.359377 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.753852 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8423c766-29a0-455c-a843-554ed1b96c71" path="/var/lib/kubelet/pods/8423c766-29a0-455c-a843-554ed1b96c71/volumes" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.763244 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7797b97577-5df66"] Mar 17 18:50:52 crc kubenswrapper[5110]: W0317 18:50:52.774109 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47daee45_2558_42ed_a85c_a2a4f350b922.slice/crio-56786ae80e5a112791138b15261f57ba7d33af925a8a6d751ef387269deb1f3e WatchSource:0}: Error finding container 56786ae80e5a112791138b15261f57ba7d33af925a8a6d751ef387269deb1f3e: Status 404 returned error can't find the container with id 56786ae80e5a112791138b15261f57ba7d33af925a8a6d751ef387269deb1f3e Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.830685 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" event={"ID":"853aa33a-fe6b-4972-a94b-077ab373a5b2","Type":"ContainerDied","Data":"8b2f81d7c2393b2ecfc4d38cb9f759ccd30145e8b805f0bebd21fd26b57ae087"} Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.830741 5110 scope.go:117] "RemoveContainer" containerID="ba4d082837844f2568db307913ca6c85a31c2f4959ebd5cc09dff086bc4b5f8f" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.830937 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77478b5f5-lvvct" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.833828 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" event={"ID":"47daee45-2558-42ed-a85c-a2a4f350b922","Type":"ContainerStarted","Data":"56786ae80e5a112791138b15261f57ba7d33af925a8a6d751ef387269deb1f3e"} Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.836855 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" event={"ID":"a54ef47e-80e3-4de6-855b-45657e28c18c","Type":"ContainerStarted","Data":"7b6feb308171cacb0f18182db3e7898d6ed4ad26739fb47971632e73c65037d1"} Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.837208 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.837256 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" event={"ID":"a54ef47e-80e3-4de6-855b-45657e28c18c","Type":"ContainerStarted","Data":"727d20e5518079dd4e0bbcefc9be316138f46e20999ea9bc1922f09589430893"} Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.859208 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" podStartSLOduration=1.859186819 podStartE2EDuration="1.859186819s" podCreationTimestamp="2026-03-17 18:50:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:52.855773874 +0000 UTC m=+196.888385396" watchObservedRunningTime="2026-03-17 18:50:52.859186819 +0000 UTC m=+196.891798361" Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.867778 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77478b5f5-lvvct"] Mar 17 18:50:52 crc kubenswrapper[5110]: I0317 18:50:52.871295 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-77478b5f5-lvvct"] Mar 17 18:50:53 crc kubenswrapper[5110]: I0317 18:50:53.110601 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6bf65448ff-xfwd6" Mar 17 18:50:53 crc kubenswrapper[5110]: I0317 18:50:53.854330 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" event={"ID":"47daee45-2558-42ed-a85c-a2a4f350b922","Type":"ContainerStarted","Data":"e360af434b199516cbcc814f607218bae8b370ec389fc27862438b776746d8ec"} Mar 17 18:50:53 crc kubenswrapper[5110]: I0317 18:50:53.855074 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:53 crc kubenswrapper[5110]: I0317 18:50:53.864502 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" Mar 17 18:50:53 crc kubenswrapper[5110]: I0317 18:50:53.880982 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7797b97577-5df66" podStartSLOduration=2.8809586190000003 podStartE2EDuration="2.880958619s" podCreationTimestamp="2026-03-17 18:50:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:50:53.878747287 +0000 UTC m=+197.911358809" watchObservedRunningTime="2026-03-17 18:50:53.880958619 +0000 UTC m=+197.913570151" Mar 17 18:50:54 crc kubenswrapper[5110]: I0317 18:50:54.761231 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="853aa33a-fe6b-4972-a94b-077ab373a5b2" path="/var/lib/kubelet/pods/853aa33a-fe6b-4972-a94b-077ab373a5b2/volumes" Mar 17 18:51:03 crc kubenswrapper[5110]: I0317 18:51:03.861372 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" containerName="oauth-openshift" containerID="cri-o://aea08505f36332454deb8437feb446a7b205125aaac2f39f4cda929bcf73116c" gracePeriod=15 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.728601 5110 kubelet.go:2547] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.729525 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" containerID="cri-o://2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f" gracePeriod=15 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.729709 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" containerID="cri-o://a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48" gracePeriod=15 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.729765 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b" gracePeriod=15 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.729810 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02" gracePeriod=15 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.729854 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" containerID="cri-o://9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5" gracePeriod=15 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.730935 5110 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731511 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731524 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731533 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731538 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731547 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731553 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731562 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="setup" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731566 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="setup" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731578 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731584 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731593 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731600 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731606 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731611 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731627 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731632 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731640 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731646 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731743 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731752 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731760 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731768 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731778 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731786 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731793 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731799 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731904 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731911 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.731988 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.739450 5110 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.748334 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.755463 5110 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="3a14caf222afb62aaabdc47808b6f944" podUID="57755cc5f99000cc11e193051474d4e2" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.827657 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.827713 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.827757 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.827926 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.828068 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.828153 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.828178 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.828210 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.828270 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.828297 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.900399 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: E0317 18:51:04.901977 5110 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.918128 5110 generic.go:358] "Generic (PLEG): container finished" podID="c2d80dcc-dd1a-434c-885d-16a7db229591" containerID="aea08505f36332454deb8437feb446a7b205125aaac2f39f4cda929bcf73116c" exitCode=0 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.918243 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" event={"ID":"c2d80dcc-dd1a-434c-885d-16a7db229591","Type":"ContainerDied","Data":"aea08505f36332454deb8437feb446a7b205125aaac2f39f4cda929bcf73116c"} Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.921382 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.923128 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.924076 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48" exitCode=0 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.924106 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b" exitCode=0 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.924116 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02" exitCode=0 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.924125 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5" exitCode=2 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.924172 5110 scope.go:117] "RemoveContainer" containerID="ad834800453a4378cd3a78013aba32739d942556cb3a1ddbecd72c90cf6f9220" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.929870 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.929938 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.929957 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.929975 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930004 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930022 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930071 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930091 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930110 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930195 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930228 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930279 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930461 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930492 5110 generic.go:358] "Generic (PLEG): container finished" podID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" containerID="7af855336ae1f74151b8d1db4c8cb8eaed795808011b6bbde668a0b81c5535a5" exitCode=0 Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930553 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7ecce33d-988f-4c1d-b3ee-37c9e949a3df","Type":"ContainerDied","Data":"7af855336ae1f74151b8d1db4c8cb8eaed795808011b6bbde668a0b81c5535a5"} Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930621 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930666 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930694 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930722 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.930751 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.931221 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.931289 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:04 crc kubenswrapper[5110]: I0317 18:51:04.931339 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.193562 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.194445 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.194980 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.203710 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:05 crc kubenswrapper[5110]: E0317 18:51:05.227019 5110 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189db5919ad4ab79 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f7dbc7e1ee9c187a863ef9b473fad27b,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:51:05.226357625 +0000 UTC m=+209.258969157,LastTimestamp:2026-03-17 18:51:05.226357625 +0000 UTC m=+209.258969157,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235424 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-provider-selection\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235496 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-error\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235634 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-policies\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235689 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-trusted-ca-bundle\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235736 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm7qv\" (UniqueName: \"kubernetes.io/projected/c2d80dcc-dd1a-434c-885d-16a7db229591-kube-api-access-vm7qv\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235817 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-cliconfig\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235882 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-idp-0-file-data\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235914 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-service-ca\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.235959 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-serving-cert\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236020 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-router-certs\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236102 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-ocp-branding-template\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236137 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-dir\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236174 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-session\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236227 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-login\") pod \"c2d80dcc-dd1a-434c-885d-16a7db229591\" (UID: \"c2d80dcc-dd1a-434c-885d-16a7db229591\") " Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236338 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236519 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236567 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236689 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236842 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236866 5110 reconciler_common.go:299] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236879 5110 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.236893 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.237399 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.240847 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.240910 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d80dcc-dd1a-434c-885d-16a7db229591-kube-api-access-vm7qv" (OuterVolumeSpecName: "kube-api-access-vm7qv") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "kube-api-access-vm7qv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.241803 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.241829 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.242006 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.242347 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.242566 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.243013 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.243110 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c2d80dcc-dd1a-434c-885d-16a7db229591" (UID: "c2d80dcc-dd1a-434c-885d-16a7db229591"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338100 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338141 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vm7qv\" (UniqueName: \"kubernetes.io/projected/c2d80dcc-dd1a-434c-885d-16a7db229591-kube-api-access-vm7qv\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338153 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338163 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338172 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338180 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338189 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338200 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338209 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.338218 5110 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2d80dcc-dd1a-434c-885d-16a7db229591-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.943037 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.943092 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" event={"ID":"c2d80dcc-dd1a-434c-885d-16a7db229591","Type":"ContainerDied","Data":"9709f9ca1616996feb5d65ff651a92b74e19640ae59c1c31ccb434e0d38c0412"} Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.943734 5110 scope.go:117] "RemoveContainer" containerID="aea08505f36332454deb8437feb446a7b205125aaac2f39f4cda929bcf73116c" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.946580 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.948046 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.964258 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.967866 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f7dbc7e1ee9c187a863ef9b473fad27b","Type":"ContainerStarted","Data":"71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e"} Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.967922 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f7dbc7e1ee9c187a863ef9b473fad27b","Type":"ContainerStarted","Data":"e4ea9c4085e9f36fbf272efa713c640f71b459cb4706d91c3e33f4bffed4083a"} Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.968456 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.970073 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: E0317 18:51:05.970105 5110 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.970370 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.974218 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:05 crc kubenswrapper[5110]: I0317 18:51:05.974638 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.314233 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.315238 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.315731 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.453760 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-var-lock\") pod \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.454278 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kubelet-dir\") pod \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.453891 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-var-lock" (OuterVolumeSpecName: "var-lock") pod "7ecce33d-988f-4c1d-b3ee-37c9e949a3df" (UID: "7ecce33d-988f-4c1d-b3ee-37c9e949a3df"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.454337 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7ecce33d-988f-4c1d-b3ee-37c9e949a3df" (UID: "7ecce33d-988f-4c1d-b3ee-37c9e949a3df"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.454828 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kube-api-access\") pod \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\" (UID: \"7ecce33d-988f-4c1d-b3ee-37c9e949a3df\") " Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.455484 5110 reconciler_common.go:299] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-var-lock\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.455645 5110 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.461270 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7ecce33d-988f-4c1d-b3ee-37c9e949a3df" (UID: "7ecce33d-988f-4c1d-b3ee-37c9e949a3df"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.557875 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ecce33d-988f-4c1d-b3ee-37c9e949a3df-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.749574 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.750534 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.977223 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7ecce33d-988f-4c1d-b3ee-37c9e949a3df","Type":"ContainerDied","Data":"b18d32500bcdc3cc6a9dc085ee1037348503ea6e8df5671d11469b3eccb5dcb5"} Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.981068 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b18d32500bcdc3cc6a9dc085ee1037348503ea6e8df5671d11469b3eccb5dcb5" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.977525 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.985382 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:06 crc kubenswrapper[5110]: I0317 18:51:06.985933 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.216343 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.217889 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.218748 5110 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.219212 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.219744 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269208 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269352 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269404 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269545 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269595 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269640 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269659 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269678 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269914 5110 reconciler_common.go:299] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269929 5110 reconciler_common.go:299] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.269941 5110 reconciler_common.go:299] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.270170 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir" (OuterVolumeSpecName: "ca-bundle-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "ca-bundle-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.271894 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.371787 5110 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.371927 5110 reconciler_common.go:299] "Volume detached for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.985694 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.988354 5110 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f" exitCode=0 Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.988611 5110 scope.go:117] "RemoveContainer" containerID="a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48" Mar 17 18:51:07 crc kubenswrapper[5110]: I0317 18:51:07.988752 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.005830 5110 scope.go:117] "RemoveContainer" containerID="ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.007125 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.007456 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.007634 5110 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.024705 5110 scope.go:117] "RemoveContainer" containerID="31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.039609 5110 scope.go:117] "RemoveContainer" containerID="9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.057906 5110 scope.go:117] "RemoveContainer" containerID="2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.072623 5110 scope.go:117] "RemoveContainer" containerID="b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.122832 5110 scope.go:117] "RemoveContainer" containerID="a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.123332 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48\": container with ID starting with a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48 not found: ID does not exist" containerID="a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.123360 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48"} err="failed to get container status \"a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48\": rpc error: code = NotFound desc = could not find container \"a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48\": container with ID starting with a787da854b2cdf85f1866cdb12db53b8f2963c8b8b23119b46b633d0e3062c48 not found: ID does not exist" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.123408 5110 scope.go:117] "RemoveContainer" containerID="ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.123566 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b\": container with ID starting with ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b not found: ID does not exist" containerID="ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.123585 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b"} err="failed to get container status \"ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b\": rpc error: code = NotFound desc = could not find container \"ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b\": container with ID starting with ce60a70341d0276e7e0edcd09551dc4f4be5f0d76ca6299faee19ca5a113777b not found: ID does not exist" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.123601 5110 scope.go:117] "RemoveContainer" containerID="31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.123774 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02\": container with ID starting with 31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02 not found: ID does not exist" containerID="31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.123798 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02"} err="failed to get container status \"31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02\": rpc error: code = NotFound desc = could not find container \"31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02\": container with ID starting with 31abced150fdf2b1fae2f237c6a3af264f7856c2a80d6abbc6fa7e42424efb02 not found: ID does not exist" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.123843 5110 scope.go:117] "RemoveContainer" containerID="9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.124025 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5\": container with ID starting with 9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5 not found: ID does not exist" containerID="9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.124043 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5"} err="failed to get container status \"9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5\": rpc error: code = NotFound desc = could not find container \"9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5\": container with ID starting with 9b6d0d649358392f8402d02e7a47e16e06cd45d92eebc1dceb786826d74472f5 not found: ID does not exist" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.124119 5110 scope.go:117] "RemoveContainer" containerID="2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.124301 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f\": container with ID starting with 2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f not found: ID does not exist" containerID="2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.124320 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f"} err="failed to get container status \"2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f\": rpc error: code = NotFound desc = could not find container \"2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f\": container with ID starting with 2d60d445baca7c53024c35dc758194e860e086c2c44a1f112c3e63fe7eecd94f not found: ID does not exist" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.124334 5110 scope.go:117] "RemoveContainer" containerID="b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.124508 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\": container with ID starting with b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520 not found: ID does not exist" containerID="b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.124527 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520"} err="failed to get container status \"b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\": rpc error: code = NotFound desc = could not find container \"b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520\": container with ID starting with b7a2e0fbcdbe11dade510304249dd42ce25f8906c84c2e725a762d8621268520 not found: ID does not exist" Mar 17 18:51:08 crc kubenswrapper[5110]: E0317 18:51:08.460630 5110 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189db5919ad4ab79 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f7dbc7e1ee9c187a863ef9b473fad27b,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:51:05.226357625 +0000 UTC m=+209.258969157,LastTimestamp:2026-03-17 18:51:05.226357625 +0000 UTC m=+209.258969157,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:51:08 crc kubenswrapper[5110]: I0317 18:51:08.757524 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a14caf222afb62aaabdc47808b6f944" path="/var/lib/kubelet/pods/3a14caf222afb62aaabdc47808b6f944/volumes" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.585172 5110 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.586151 5110 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.586607 5110 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.587083 5110 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.587607 5110 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:10 crc kubenswrapper[5110]: I0317 18:51:10.587641 5110 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.587940 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="200ms" Mar 17 18:51:10 crc kubenswrapper[5110]: E0317 18:51:10.789363 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="400ms" Mar 17 18:51:11 crc kubenswrapper[5110]: E0317 18:51:11.190922 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="800ms" Mar 17 18:51:11 crc kubenswrapper[5110]: E0317 18:51:11.991806 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="1.6s" Mar 17 18:51:12 crc kubenswrapper[5110]: I0317 18:51:12.133333 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:51:12 crc kubenswrapper[5110]: I0317 18:51:12.133757 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:51:13 crc kubenswrapper[5110]: E0317 18:51:13.593590 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="3.2s" Mar 17 18:51:16 crc kubenswrapper[5110]: I0317 18:51:16.749531 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:16 crc kubenswrapper[5110]: I0317 18:51:16.750463 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:16 crc kubenswrapper[5110]: E0317 18:51:16.794998 5110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.230:6443: connect: connection refused" interval="6.4s" Mar 17 18:51:18 crc kubenswrapper[5110]: E0317 18:51:18.461851 5110 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.230:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189db5919ad4ab79 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f7dbc7e1ee9c187a863ef9b473fad27b,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 18:51:05.226357625 +0000 UTC m=+209.258969157,LastTimestamp:2026-03-17 18:51:05.226357625 +0000 UTC m=+209.258969157,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 18:51:18 crc kubenswrapper[5110]: I0317 18:51:18.744730 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:18 crc kubenswrapper[5110]: I0317 18:51:18.745615 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:18 crc kubenswrapper[5110]: I0317 18:51:18.746189 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:18 crc kubenswrapper[5110]: I0317 18:51:18.759233 5110 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:18 crc kubenswrapper[5110]: I0317 18:51:18.759270 5110 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:18 crc kubenswrapper[5110]: E0317 18:51:18.759809 5110 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:18 crc kubenswrapper[5110]: I0317 18:51:18.760051 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:18 crc kubenswrapper[5110]: E0317 18:51:18.829908 5110 desired_state_of_world_populator.go:305] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" volumeName="registry-storage" Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.058400 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"2a4139779d022d87dad08f6924ab255a662628bb576258d67d8019803d04eaf4"} Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.061279 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.061326 5110 generic.go:358] "Generic (PLEG): container finished" podID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerID="e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3" exitCode=1 Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.061454 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerDied","Data":"e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3"} Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.062155 5110 scope.go:117] "RemoveContainer" containerID="e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3" Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.062457 5110 status_manager.go:895] "Failed to get status for pod" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.062946 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.063180 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:19 crc kubenswrapper[5110]: I0317 18:51:19.160446 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.069071 5110 generic.go:358] "Generic (PLEG): container finished" podID="57755cc5f99000cc11e193051474d4e2" containerID="bb4a378f3bf3376a3eb0a44a6ef1fc284699591e6ee880918ccfffc8f7dac664" exitCode=0 Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.070241 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.069146 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerDied","Data":"bb4a378f3bf3376a3eb0a44a6ef1fc284699591e6ee880918ccfffc8f7dac664"} Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.069612 5110 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.070430 5110 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.070797 5110 status_manager.go:895] "Failed to get status for pod" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:20 crc kubenswrapper[5110]: E0317 18:51:20.070862 5110 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.071216 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.075155 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.075403 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"a526d5facb1dd620a768235981acef08fe9c674d5a28c481489ad7e454766827"} Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.076389 5110 status_manager.go:895] "Failed to get status for pod" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.076885 5110 status_manager.go:895] "Failed to get status for pod" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:20 crc kubenswrapper[5110]: I0317 18:51:20.077384 5110 status_manager.go:895] "Failed to get status for pod" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" pod="openshift-authentication/oauth-openshift-66458b6674-zpddx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-zpddx\": dial tcp 38.102.83.230:6443: connect: connection refused" Mar 17 18:51:21 crc kubenswrapper[5110]: I0317 18:51:21.089252 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"0f2ed4101edbfa07becf7e33485df5942067ee3e4e1c07c02d25722c4ab15231"} Mar 17 18:51:21 crc kubenswrapper[5110]: I0317 18:51:21.089815 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"744cdde592017315312d05f1ea0cd708127de4651d1ff3e09b973263327618ff"} Mar 17 18:51:21 crc kubenswrapper[5110]: I0317 18:51:21.089828 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"bf526c8a98fb402dbe7fb98ee5ccbf1eb5041c7b8ff41b060c38b89052e4af19"} Mar 17 18:51:21 crc kubenswrapper[5110]: I0317 18:51:21.686068 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:51:21 crc kubenswrapper[5110]: I0317 18:51:21.686210 5110 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 17 18:51:21 crc kubenswrapper[5110]: I0317 18:51:21.686555 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 17 18:51:22 crc kubenswrapper[5110]: I0317 18:51:22.096954 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"24ea7760a3aca114eb87b7db1c1dcc4fceaeeb8bf698925ce02eb391f47a756f"} Mar 17 18:51:22 crc kubenswrapper[5110]: I0317 18:51:22.097001 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"828edf40b1e53d458dcd4e93276b2f546ea06a470cca7eafdfbab5245b78bad2"} Mar 17 18:51:22 crc kubenswrapper[5110]: I0317 18:51:22.097478 5110 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:22 crc kubenswrapper[5110]: I0317 18:51:22.097492 5110 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:23 crc kubenswrapper[5110]: I0317 18:51:23.761096 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:23 crc kubenswrapper[5110]: I0317 18:51:23.761544 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:23 crc kubenswrapper[5110]: I0317 18:51:23.771697 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:26 crc kubenswrapper[5110]: I0317 18:51:26.904900 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:51:27 crc kubenswrapper[5110]: I0317 18:51:27.222684 5110 kubelet.go:3329] "Deleted mirror pod as it didn't match the static Pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:27 crc kubenswrapper[5110]: I0317 18:51:27.222742 5110 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:27 crc kubenswrapper[5110]: I0317 18:51:27.376187 5110 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="4b982fbc-c4a0-446b-ad92-71e54438b983" Mar 17 18:51:28 crc kubenswrapper[5110]: I0317 18:51:28.135559 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:28 crc kubenswrapper[5110]: I0317 18:51:28.135609 5110 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:28 crc kubenswrapper[5110]: I0317 18:51:28.136098 5110 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:28 crc kubenswrapper[5110]: I0317 18:51:28.139759 5110 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="4b982fbc-c4a0-446b-ad92-71e54438b983" Mar 17 18:51:29 crc kubenswrapper[5110]: I0317 18:51:29.140585 5110 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:29 crc kubenswrapper[5110]: I0317 18:51:29.141215 5110 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:29 crc kubenswrapper[5110]: I0317 18:51:29.144662 5110 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="4b982fbc-c4a0-446b-ad92-71e54438b983" Mar 17 18:51:29 crc kubenswrapper[5110]: I0317 18:51:29.148426 5110 status_manager.go:346] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://bf526c8a98fb402dbe7fb98ee5ccbf1eb5041c7b8ff41b060c38b89052e4af19" Mar 17 18:51:29 crc kubenswrapper[5110]: I0317 18:51:29.148450 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:30 crc kubenswrapper[5110]: I0317 18:51:30.148716 5110 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:30 crc kubenswrapper[5110]: I0317 18:51:30.149334 5110 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="331f5282-47de-4f21-903e-5dfb189879f4" Mar 17 18:51:30 crc kubenswrapper[5110]: I0317 18:51:30.152647 5110 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="4b982fbc-c4a0-446b-ad92-71e54438b983" Mar 17 18:51:31 crc kubenswrapper[5110]: I0317 18:51:31.685962 5110 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 17 18:51:31 crc kubenswrapper[5110]: I0317 18:51:31.686037 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 17 18:51:37 crc kubenswrapper[5110]: I0317 18:51:37.183700 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Mar 17 18:51:37 crc kubenswrapper[5110]: I0317 18:51:37.302032 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-kknhg\"" Mar 17 18:51:37 crc kubenswrapper[5110]: I0317 18:51:37.533878 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.009307 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.293028 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"serving-cert\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.390237 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.704474 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-serving-cert\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.707930 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"etcd-serving-ca\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.730751 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"image-import-ca\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.778642 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9pgs7\"" Mar 17 18:51:38 crc kubenswrapper[5110]: I0317 18:51:38.787123 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.243258 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-tjs74\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.349632 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.362796 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"pprof-cert\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.422613 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"kube-root-ca.crt\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.461754 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.464104 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-kl6m8\"" Mar 17 18:51:39 crc kubenswrapper[5110]: I0317 18:51:39.913164 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.049747 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.080952 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"config\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.219606 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.225948 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-images\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.275765 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"kube-root-ca.crt\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.469749 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.583900 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-config\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.590267 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.643131 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-w9nzh\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.739327 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.781001 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.788789 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-scheduler-operator-serving-cert\"" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.789842 5110 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:51:40 crc kubenswrapper[5110]: I0317 18:51:40.832172 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"metrics-tls\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.016217 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-ca-bundle\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.023139 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-dockercfg-dzw6b\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.089959 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.122278 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.174639 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.186168 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.257346 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bgxvm\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.282508 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-4zqgh\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.296112 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.302687 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.346473 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.356283 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-dockercfg-gnx66\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.407350 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.487685 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serviceaccount-dockercfg-4gqzj\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.488990 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.507508 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serving-cert\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.620144 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.686043 5110 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.686191 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.686279 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.687494 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"a526d5facb1dd620a768235981acef08fe9c674d5a28c481489ad7e454766827"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.687726 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="kube-controller-manager" containerID="cri-o://a526d5facb1dd620a768235981acef08fe9c674d5a28c481489ad7e454766827" gracePeriod=30 Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.758852 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"service-ca-bundle\"" Mar 17 18:51:41 crc kubenswrapper[5110]: I0317 18:51:41.956643 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.087706 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-serving-cert\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.094408 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-l2v2m\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.133200 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.133273 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.199310 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-dockercfg-sw6nc\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.216864 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.236625 5110 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.241863 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-66458b6674-zpddx"] Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.241923 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.247189 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.260702 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.260685603 podStartE2EDuration="15.260685603s" podCreationTimestamp="2026-03-17 18:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:51:42.257559736 +0000 UTC m=+246.290171258" watchObservedRunningTime="2026-03-17 18:51:42.260685603 +0000 UTC m=+246.293297125" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.265543 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.330725 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"dns-operator-dockercfg-wbbsn\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.398583 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.417012 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-tls\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.581276 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-node-identity\"/\"network-node-identity-cert\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.629504 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.756317 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mcc-proxy-tls\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.758044 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" path="/var/lib/kubelet/pods/c2d80dcc-dd1a-434c-885d-16a7db229591/volumes" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.786232 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8dkm8\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.794190 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.847469 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-admission-controller-secret\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.872813 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-global-ca\"" Mar 17 18:51:42 crc kubenswrapper[5110]: I0317 18:51:42.962902 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.031465 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.082102 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.181332 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.250918 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"config\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.313966 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-dockercfg-2wbn2\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.388728 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-6tbpn\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.415999 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.469493 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"openshift-service-ca.crt\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.469844 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-g6kgg\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.485678 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.533640 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-client\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.568517 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.584938 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.667706 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.671505 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.768982 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-mmcpt\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.779339 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Mar 17 18:51:43 crc kubenswrapper[5110]: I0317 18:51:43.857649 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-6c46w\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.005702 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.035621 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"catalog-operator-serving-cert\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.070637 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.229622 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-6n5ln\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.271104 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-images\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.329135 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.458219 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-dockercfg-4vdnc\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.489574 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.516649 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.536586 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.623548 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"trusted-ca\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.758860 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.769185 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-config\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.964673 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.998830 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:44 crc kubenswrapper[5110]: I0317 18:51:44.999023 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.196030 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.311919 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.313409 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.378786 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-tls\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.407926 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.481533 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.717894 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.779603 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.807841 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.822526 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"audit-1\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.991043 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"config-operator-serving-cert\"" Mar 17 18:51:45 crc kubenswrapper[5110]: I0317 18:51:45.996741 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.032492 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.053503 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"serving-cert\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.091166 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-sjn6s\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.213248 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"env-overrides\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.254406 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.261499 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"etcd-client\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.296915 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.325361 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.341521 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.342261 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"ovnkube-identity-cm\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.405179 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.411631 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.422112 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ac-dockercfg-gj7jx\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.446517 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-controller-dockercfg-xnj77\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.478581 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"trusted-ca-bundle\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.569214 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.720263 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.821987 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.845431 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-config\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.864955 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.891203 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.911985 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-client\"" Mar 17 18:51:46 crc kubenswrapper[5110]: I0317 18:51:46.969224 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.004744 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.016518 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.051665 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.112088 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"metrics-tls\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.130787 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.191554 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"openshift-service-ca.crt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.208017 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"proxy-tls\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.292600 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-certs-default\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.297139 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"openshift-service-ca.crt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.307103 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.439642 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.475992 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-tk7bt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.548438 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.741287 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"serving-cert\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.805434 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"hostpath-provisioner\"/\"csi-hostpath-provisioner-sa-dockercfg-7dcws\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.844626 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"metrics-tls\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.877878 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"encryption-config-1\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.904420 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.908085 5110 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.919943 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\"" Mar 17 18:51:47 crc kubenswrapper[5110]: I0317 18:51:47.927292 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"client-ca\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.043489 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.063530 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.064686 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.066853 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.124554 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.169631 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-kpvmz\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.259247 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.274388 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.289279 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.339529 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.351693 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"kube-root-ca.crt\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.451210 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-operator-tls\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.579670 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-metrics\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.644415 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.697014 5110 kubelet.go:2547] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.697374 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" containerID="cri-o://71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e" gracePeriod=5 Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.716971 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\"" Mar 17 18:51:48 crc kubenswrapper[5110]: I0317 18:51:48.905045 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"cluster-image-registry-operator-dockercfg-ntnd7\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.034291 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.062808 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"packageserver-service-cert\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.160816 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.176196 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.293903 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.308141 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-root-ca.crt\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.315264 5110 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.479535 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-nl8tp\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.502711 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-bjqfd\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.523402 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-kw8fx\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.524674 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-config\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.562828 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"client-ca\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.619679 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-dockercfg-2cfkp\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.630947 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"serving-cert\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.662357 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-nwglk\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.701144 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.759857 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-serving-cert\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.772536 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"serving-cert\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.794569 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.797674 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-service-ca-bundle\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.837913 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.899789 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Mar 17 18:51:49 crc kubenswrapper[5110]: I0317 18:51:49.901871 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.005648 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mco-proxy-tls\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.172112 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6468d5b88f-mzmx6"] Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173664 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" containerName="installer" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173699 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" containerName="installer" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173744 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" containerName="oauth-openshift" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173752 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" containerName="oauth-openshift" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173762 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173768 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173895 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173912 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="7ecce33d-988f-4c1d-b3ee-37c9e949a3df" containerName="installer" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.173923 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="c2d80dcc-dd1a-434c-885d-16a7db229591" containerName="oauth-openshift" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.178932 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.190276 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.190226 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.194673 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.209264 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.210004 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.211250 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-session\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.212824 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"openshift-service-ca.crt\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.213391 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.213878 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-d2bf2\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.215047 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"kube-root-ca.crt\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.215237 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.216407 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.217579 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"audit\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.219876 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.220715 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.232465 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6468d5b88f-mzmx6"] Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.235658 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.322729 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.322973 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-service-ca\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.323160 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-session\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.323329 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.323514 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.323673 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.323840 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-error\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.323998 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/844e74db-f5b2-4ae5-bd51-b8960f1bf302-audit-dir\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.324175 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.324328 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2pxt\" (UniqueName: \"kubernetes.io/projected/844e74db-f5b2-4ae5-bd51-b8960f1bf302-kube-api-access-g2pxt\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.324478 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-router-certs\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.324651 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-login\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.324804 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-audit-policies\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.324953 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.358467 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.426711 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/844e74db-f5b2-4ae5-bd51-b8960f1bf302-audit-dir\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.426779 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.426822 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/844e74db-f5b2-4ae5-bd51-b8960f1bf302-audit-dir\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.426908 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g2pxt\" (UniqueName: \"kubernetes.io/projected/844e74db-f5b2-4ae5-bd51-b8960f1bf302-kube-api-access-g2pxt\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.426968 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-router-certs\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.427047 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-login\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.427173 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-audit-policies\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.427455 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.427608 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.427742 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-service-ca\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.427842 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-session\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.428121 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.428261 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.428338 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.428402 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-error\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.428647 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-audit-policies\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.429043 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-service-ca\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.429352 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.430273 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.433356 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.434485 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-router-certs\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.435454 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-session\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.435528 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.450549 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-login\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.453014 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.467210 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-user-template-error\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.469028 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/844e74db-f5b2-4ae5-bd51-b8960f1bf302-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.469726 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-jcmfj\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.480858 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-dockercfg-bf7fj\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.498755 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2pxt\" (UniqueName: \"kubernetes.io/projected/844e74db-f5b2-4ae5-bd51-b8960f1bf302-kube-api-access-g2pxt\") pod \"oauth-openshift-6468d5b88f-mzmx6\" (UID: \"844e74db-f5b2-4ae5-bd51-b8960f1bf302\") " pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.499197 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.562416 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.563577 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"package-server-manager-serving-cert\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.568420 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-dockercfg-tnfx9\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.569393 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"marketplace-trusted-ca\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.610337 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.703586 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-mdwwj\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.806735 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.834024 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-tls\"" Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.942094 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6468d5b88f-mzmx6"] Mar 17 18:51:50 crc kubenswrapper[5110]: I0317 18:51:50.980214 5110 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:50.999929 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"audit-1\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.054164 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"node-bootstrapper-token\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.056607 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.057380 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.083109 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.090150 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.110742 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.131822 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.140682 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"config\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.194616 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.287458 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" event={"ID":"844e74db-f5b2-4ae5-bd51-b8960f1bf302","Type":"ContainerStarted","Data":"1df81a6232c7db32e7dec65788bf2fe097d083a188ccb1891b00fe33b9871367"} Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.287501 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" event={"ID":"844e74db-f5b2-4ae5-bd51-b8960f1bf302","Type":"ContainerStarted","Data":"63ab1602c97b5ff64785cee98783c19647af7e296fadb240d6c13fd524478b02"} Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.287888 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.312653 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" podStartSLOduration=73.312632346 podStartE2EDuration="1m13.312632346s" podCreationTimestamp="2026-03-17 18:50:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:51:51.303991965 +0000 UTC m=+255.336603497" watchObservedRunningTime="2026-03-17 18:51:51.312632346 +0000 UTC m=+255.345243868" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.366427 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.403820 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-2h6bs\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.563868 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.683026 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.708138 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.708545 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-jmhxf\"" Mar 17 18:51:51 crc kubenswrapper[5110]: I0317 18:51:51.858597 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-config\"" Mar 17 18:51:52 crc kubenswrapper[5110]: I0317 18:51:52.034812 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Mar 17 18:51:52 crc kubenswrapper[5110]: I0317 18:51:52.103604 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Mar 17 18:51:52 crc kubenswrapper[5110]: I0317 18:51:52.281255 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6468d5b88f-mzmx6" Mar 17 18:51:52 crc kubenswrapper[5110]: I0317 18:51:52.296855 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 17 18:51:52 crc kubenswrapper[5110]: I0317 18:51:52.718797 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-qqw4z\"" Mar 17 18:51:52 crc kubenswrapper[5110]: I0317 18:51:52.774444 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-djmfg\"" Mar 17 18:51:53 crc kubenswrapper[5110]: I0317 18:51:53.048179 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-rbac-proxy\"" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.302994 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.303504 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.305958 5110 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.308836 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.308908 5110 generic.go:358] "Generic (PLEG): container finished" podID="f7dbc7e1ee9c187a863ef9b473fad27b" containerID="71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e" exitCode=137 Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.309018 5110 scope.go:117] "RemoveContainer" containerID="71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.309046 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.333236 5110 scope.go:117] "RemoveContainer" containerID="71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e" Mar 17 18:51:54 crc kubenswrapper[5110]: E0317 18:51:54.333667 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e\": container with ID starting with 71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e not found: ID does not exist" containerID="71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.333725 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e"} err="failed to get container status \"71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e\": rpc error: code = NotFound desc = could not find container \"71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e\": container with ID starting with 71ab1600332032a292cfc0c6e5c454421a1fce2e94d18e9db2fe183a1d87884e not found: ID does not exist" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424392 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424507 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424535 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424592 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424627 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424646 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock" (OuterVolumeSpecName: "var-lock") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424702 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests" (OuterVolumeSpecName: "manifests") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424730 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.424787 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log" (OuterVolumeSpecName: "var-log") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.425461 5110 reconciler_common.go:299] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.425502 5110 reconciler_common.go:299] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.425521 5110 reconciler_common.go:299] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.425538 5110 reconciler_common.go:299] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.439566 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.527283 5110 reconciler_common.go:299] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.640274 5110 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Mar 17 18:51:54 crc kubenswrapper[5110]: I0317 18:51:54.758617 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" path="/var/lib/kubelet/pods/f7dbc7e1ee9c187a863ef9b473fad27b/volumes" Mar 17 18:52:09 crc kubenswrapper[5110]: I0317 18:52:09.409994 5110 generic.go:358] "Generic (PLEG): container finished" podID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerID="b8e1d8bfcc1bf34ea4250ceb8b99b5a12fcc1246938f9cc7324993f3b33df7d3" exitCode=0 Mar 17 18:52:09 crc kubenswrapper[5110]: I0317 18:52:09.410838 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" event={"ID":"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5","Type":"ContainerDied","Data":"b8e1d8bfcc1bf34ea4250ceb8b99b5a12fcc1246938f9cc7324993f3b33df7d3"} Mar 17 18:52:09 crc kubenswrapper[5110]: I0317 18:52:09.411282 5110 scope.go:117] "RemoveContainer" containerID="b8e1d8bfcc1bf34ea4250ceb8b99b5a12fcc1246938f9cc7324993f3b33df7d3" Mar 17 18:52:10 crc kubenswrapper[5110]: I0317 18:52:10.420791 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" event={"ID":"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5","Type":"ContainerStarted","Data":"d5ce783b9323ff971b5b6c0469ad3e4208d7689e5474a70f8f41e7093fdf50d8"} Mar 17 18:52:10 crc kubenswrapper[5110]: I0317 18:52:10.422335 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:52:10 crc kubenswrapper[5110]: I0317 18:52:10.424705 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:52:11 crc kubenswrapper[5110]: I0317 18:52:11.277115 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-wzhvk\"" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.132485 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.133108 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.133203 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.134394 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0075007a0a6d1dcc16449135cf348101cd8615648447feb043c5978c7d129075"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.134507 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://0075007a0a6d1dcc16449135cf348101cd8615648447feb043c5978c7d129075" gracePeriod=600 Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.439432 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.445618 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.445819 5110 generic.go:358] "Generic (PLEG): container finished" podID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerID="a526d5facb1dd620a768235981acef08fe9c674d5a28c481489ad7e454766827" exitCode=137 Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.445920 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerDied","Data":"a526d5facb1dd620a768235981acef08fe9c674d5a28c481489ad7e454766827"} Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.445989 5110 scope.go:117] "RemoveContainer" containerID="e953ac8aba754ae36281af789c97a87f507f56de91222c1c48f06aa9b4d087e3" Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.451690 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="0075007a0a6d1dcc16449135cf348101cd8615648447feb043c5978c7d129075" exitCode=0 Mar 17 18:52:12 crc kubenswrapper[5110]: I0317 18:52:12.451740 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"0075007a0a6d1dcc16449135cf348101cd8615648447feb043c5978c7d129075"} Mar 17 18:52:13 crc kubenswrapper[5110]: I0317 18:52:13.462128 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"ebe54d3668f3ba6d72d2f6801190bcc6cdfdce7f9ffcdbc79df03266e49b4ac1"} Mar 17 18:52:13 crc kubenswrapper[5110]: I0317 18:52:13.466457 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 18:52:13 crc kubenswrapper[5110]: I0317 18:52:13.468531 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"01a07e90b3ff718e2fde29a82f0d748381dec80fcce9f63c34ed9b5b653f366e"} Mar 17 18:52:13 crc kubenswrapper[5110]: I0317 18:52:13.899001 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"default-dockercfg-hqpm5\"" Mar 17 18:52:16 crc kubenswrapper[5110]: I0317 18:52:16.905508 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:52:21 crc kubenswrapper[5110]: I0317 18:52:21.685546 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:52:21 crc kubenswrapper[5110]: I0317 18:52:21.690152 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:52:22 crc kubenswrapper[5110]: I0317 18:52:22.538880 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 18:52:28 crc kubenswrapper[5110]: I0317 18:52:28.781926 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"ingress-operator-dockercfg-74nwh\"" Mar 17 18:52:31 crc kubenswrapper[5110]: I0317 18:52:31.983913 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.104223 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562892-2hdnb"] Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.114499 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.115599 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562892-2hdnb"] Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.116308 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.116413 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.119317 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.170887 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxt6h\" (UniqueName: \"kubernetes.io/projected/64571774-f18c-4140-b48c-f8868af19eb7-kube-api-access-rxt6h\") pod \"auto-csr-approver-29562892-2hdnb\" (UID: \"64571774-f18c-4140-b48c-f8868af19eb7\") " pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.271546 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rxt6h\" (UniqueName: \"kubernetes.io/projected/64571774-f18c-4140-b48c-f8868af19eb7-kube-api-access-rxt6h\") pod \"auto-csr-approver-29562892-2hdnb\" (UID: \"64571774-f18c-4140-b48c-f8868af19eb7\") " pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.314924 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxt6h\" (UniqueName: \"kubernetes.io/projected/64571774-f18c-4140-b48c-f8868af19eb7-kube-api-access-rxt6h\") pod \"auto-csr-approver-29562892-2hdnb\" (UID: \"64571774-f18c-4140-b48c-f8868af19eb7\") " pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.433144 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:32 crc kubenswrapper[5110]: I0317 18:52:32.831020 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562892-2hdnb"] Mar 17 18:52:33 crc kubenswrapper[5110]: I0317 18:52:33.594905 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" event={"ID":"64571774-f18c-4140-b48c-f8868af19eb7","Type":"ContainerStarted","Data":"67d9613d18652e799b1752e547b6feafc4f786395abd797efe744c044ed27170"} Mar 17 18:52:34 crc kubenswrapper[5110]: I0317 18:52:34.601275 5110 generic.go:358] "Generic (PLEG): container finished" podID="64571774-f18c-4140-b48c-f8868af19eb7" containerID="fbb91fd8cdbb67bf5e4ebf66b4eb5d69dfc2a8a7933accda111fb00479faa00b" exitCode=0 Mar 17 18:52:34 crc kubenswrapper[5110]: I0317 18:52:34.601343 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" event={"ID":"64571774-f18c-4140-b48c-f8868af19eb7","Type":"ContainerDied","Data":"fbb91fd8cdbb67bf5e4ebf66b4eb5d69dfc2a8a7933accda111fb00479faa00b"} Mar 17 18:52:35 crc kubenswrapper[5110]: I0317 18:52:35.911420 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.016575 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxt6h\" (UniqueName: \"kubernetes.io/projected/64571774-f18c-4140-b48c-f8868af19eb7-kube-api-access-rxt6h\") pod \"64571774-f18c-4140-b48c-f8868af19eb7\" (UID: \"64571774-f18c-4140-b48c-f8868af19eb7\") " Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.022624 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64571774-f18c-4140-b48c-f8868af19eb7-kube-api-access-rxt6h" (OuterVolumeSpecName: "kube-api-access-rxt6h") pod "64571774-f18c-4140-b48c-f8868af19eb7" (UID: "64571774-f18c-4140-b48c-f8868af19eb7"). InnerVolumeSpecName "kube-api-access-rxt6h". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.117486 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rxt6h\" (UniqueName: \"kubernetes.io/projected/64571774-f18c-4140-b48c-f8868af19eb7-kube-api-access-rxt6h\") on node \"crc\" DevicePath \"\"" Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.619097 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" event={"ID":"64571774-f18c-4140-b48c-f8868af19eb7","Type":"ContainerDied","Data":"67d9613d18652e799b1752e547b6feafc4f786395abd797efe744c044ed27170"} Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.619149 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67d9613d18652e799b1752e547b6feafc4f786395abd797efe744c044ed27170" Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.619237 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562892-2hdnb" Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.954540 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 18:52:36 crc kubenswrapper[5110]: I0317 18:52:36.959384 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.673797 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nldd2"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.675033 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nldd2" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="registry-server" containerID="cri-o://5ddd0af732769f3a01a5f513e450a9a6ed4a6589e2c0cb2a9bfb40f8d8b14e56" gracePeriod=30 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.681809 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-czltc"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.682222 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-czltc" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="registry-server" containerID="cri-o://052c29ca4712916b5105c422bb78da12e80470450fd5d29aeda39dd4bef5d3e6" gracePeriod=30 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.688909 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-95rbr"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.692210 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" containerID="cri-o://d5ce783b9323ff971b5b6c0469ad3e4208d7689e5474a70f8f41e7093fdf50d8" gracePeriod=30 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.702480 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh5rg"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.703109 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bh5rg" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="registry-server" containerID="cri-o://2808e465d7090bf5dc3dabbb865ad0a0bc592a6ff26bb912931ee02f7f4af634" gracePeriod=30 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.711941 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jctgs"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.712381 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jctgs" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="registry-server" containerID="cri-o://3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9" gracePeriod=30 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.725932 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-9m47z"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.727373 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="64571774-f18c-4140-b48c-f8868af19eb7" containerName="oc" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.730216 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="64571774-f18c-4140-b48c-f8868af19eb7" containerName="oc" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.735464 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="64571774-f18c-4140-b48c-f8868af19eb7" containerName="oc" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.753839 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.765553 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-9m47z"] Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.833519 5110 generic.go:358] "Generic (PLEG): container finished" podID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerID="d5ce783b9323ff971b5b6c0469ad3e4208d7689e5474a70f8f41e7093fdf50d8" exitCode=0 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.833673 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" event={"ID":"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5","Type":"ContainerDied","Data":"d5ce783b9323ff971b5b6c0469ad3e4208d7689e5474a70f8f41e7093fdf50d8"} Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.833717 5110 scope.go:117] "RemoveContainer" containerID="b8e1d8bfcc1bf34ea4250ceb8b99b5a12fcc1246938f9cc7324993f3b33df7d3" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.837688 5110 generic.go:358] "Generic (PLEG): container finished" podID="9fd9df75-65d5-476e-9e82-da35f1116996" containerID="2808e465d7090bf5dc3dabbb865ad0a0bc592a6ff26bb912931ee02f7f4af634" exitCode=0 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.837849 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh5rg" event={"ID":"9fd9df75-65d5-476e-9e82-da35f1116996","Type":"ContainerDied","Data":"2808e465d7090bf5dc3dabbb865ad0a0bc592a6ff26bb912931ee02f7f4af634"} Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.840332 5110 generic.go:358] "Generic (PLEG): container finished" podID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerID="052c29ca4712916b5105c422bb78da12e80470450fd5d29aeda39dd4bef5d3e6" exitCode=0 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.840419 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czltc" event={"ID":"a0b63bb8-f8c1-4591-affd-65e1d6cb3393","Type":"ContainerDied","Data":"052c29ca4712916b5105c422bb78da12e80470450fd5d29aeda39dd4bef5d3e6"} Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.842810 5110 generic.go:358] "Generic (PLEG): container finished" podID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerID="5ddd0af732769f3a01a5f513e450a9a6ed4a6589e2c0cb2a9bfb40f8d8b14e56" exitCode=0 Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.842904 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldd2" event={"ID":"72bdf6fb-83a5-4dcb-996c-0bee47903667","Type":"ContainerDied","Data":"5ddd0af732769f3a01a5f513e450a9a6ed4a6589e2c0cb2a9bfb40f8d8b14e56"} Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.865942 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53db1455-0000-4bb3-9f81-0e079a38a5ec-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.866211 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsmlm\" (UniqueName: \"kubernetes.io/projected/53db1455-0000-4bb3-9f81-0e079a38a5ec-kube-api-access-vsmlm\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.867434 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53db1455-0000-4bb3-9f81-0e079a38a5ec-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.867543 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/53db1455-0000-4bb3-9f81-0e079a38a5ec-tmp\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.968774 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/53db1455-0000-4bb3-9f81-0e079a38a5ec-tmp\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.969715 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53db1455-0000-4bb3-9f81-0e079a38a5ec-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.969801 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/53db1455-0000-4bb3-9f81-0e079a38a5ec-tmp\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.970242 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vsmlm\" (UniqueName: \"kubernetes.io/projected/53db1455-0000-4bb3-9f81-0e079a38a5ec-kube-api-access-vsmlm\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.970960 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53db1455-0000-4bb3-9f81-0e079a38a5ec-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.979543 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53db1455-0000-4bb3-9f81-0e079a38a5ec-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.980757 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53db1455-0000-4bb3-9f81-0e079a38a5ec-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:10 crc kubenswrapper[5110]: I0317 18:53:10.999492 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsmlm\" (UniqueName: \"kubernetes.io/projected/53db1455-0000-4bb3-9f81-0e079a38a5ec-kube-api-access-vsmlm\") pod \"marketplace-operator-547dbd544d-9m47z\" (UID: \"53db1455-0000-4bb3-9f81-0e079a38a5ec\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.127883 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.134336 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.139946 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.142233 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.143590 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.161158 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.273935 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-catalog-content\") pod \"72bdf6fb-83a5-4dcb-996c-0bee47903667\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.273976 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-utilities\") pod \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274021 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gr2k\" (UniqueName: \"kubernetes.io/projected/9fd9df75-65d5-476e-9e82-da35f1116996-kube-api-access-9gr2k\") pod \"9fd9df75-65d5-476e-9e82-da35f1116996\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274040 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-utilities\") pod \"72bdf6fb-83a5-4dcb-996c-0bee47903667\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274089 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czp6q\" (UniqueName: \"kubernetes.io/projected/72bdf6fb-83a5-4dcb-996c-0bee47903667-kube-api-access-czp6q\") pod \"72bdf6fb-83a5-4dcb-996c-0bee47903667\" (UID: \"72bdf6fb-83a5-4dcb-996c-0bee47903667\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274111 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-utilities\") pod \"9fd9df75-65d5-476e-9e82-da35f1116996\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274130 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnpvm\" (UniqueName: \"kubernetes.io/projected/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-kube-api-access-nnpvm\") pod \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274150 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4hxl\" (UniqueName: \"kubernetes.io/projected/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-kube-api-access-q4hxl\") pod \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274166 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rt98\" (UniqueName: \"kubernetes.io/projected/66594344-3d62-414a-9e89-3ab843c5239a-kube-api-access-8rt98\") pod \"66594344-3d62-414a-9e89-3ab843c5239a\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274201 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-trusted-ca\") pod \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274261 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-catalog-content\") pod \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\" (UID: \"a0b63bb8-f8c1-4591-affd-65e1d6cb3393\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274372 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-catalog-content\") pod \"66594344-3d62-414a-9e89-3ab843c5239a\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274425 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-tmp\") pod \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274453 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-utilities\") pod \"66594344-3d62-414a-9e89-3ab843c5239a\" (UID: \"66594344-3d62-414a-9e89-3ab843c5239a\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274509 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-catalog-content\") pod \"9fd9df75-65d5-476e-9e82-da35f1116996\" (UID: \"9fd9df75-65d5-476e-9e82-da35f1116996\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.274540 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-operator-metrics\") pod \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\" (UID: \"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5\") " Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.275594 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" (UID: "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.275836 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-tmp" (OuterVolumeSpecName: "tmp") pod "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" (UID: "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.275954 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-utilities" (OuterVolumeSpecName: "utilities") pod "72bdf6fb-83a5-4dcb-996c-0bee47903667" (UID: "72bdf6fb-83a5-4dcb-996c-0bee47903667"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.275954 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-utilities" (OuterVolumeSpecName: "utilities") pod "a0b63bb8-f8c1-4591-affd-65e1d6cb3393" (UID: "a0b63bb8-f8c1-4591-affd-65e1d6cb3393"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.276803 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-utilities" (OuterVolumeSpecName: "utilities") pod "66594344-3d62-414a-9e89-3ab843c5239a" (UID: "66594344-3d62-414a-9e89-3ab843c5239a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.277092 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-utilities" (OuterVolumeSpecName: "utilities") pod "9fd9df75-65d5-476e-9e82-da35f1116996" (UID: "9fd9df75-65d5-476e-9e82-da35f1116996"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.280885 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd9df75-65d5-476e-9e82-da35f1116996-kube-api-access-9gr2k" (OuterVolumeSpecName: "kube-api-access-9gr2k") pod "9fd9df75-65d5-476e-9e82-da35f1116996" (UID: "9fd9df75-65d5-476e-9e82-da35f1116996"). InnerVolumeSpecName "kube-api-access-9gr2k". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.285207 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-kube-api-access-q4hxl" (OuterVolumeSpecName: "kube-api-access-q4hxl") pod "a0b63bb8-f8c1-4591-affd-65e1d6cb3393" (UID: "a0b63bb8-f8c1-4591-affd-65e1d6cb3393"). InnerVolumeSpecName "kube-api-access-q4hxl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.285250 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72bdf6fb-83a5-4dcb-996c-0bee47903667-kube-api-access-czp6q" (OuterVolumeSpecName: "kube-api-access-czp6q") pod "72bdf6fb-83a5-4dcb-996c-0bee47903667" (UID: "72bdf6fb-83a5-4dcb-996c-0bee47903667"). InnerVolumeSpecName "kube-api-access-czp6q". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.285279 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66594344-3d62-414a-9e89-3ab843c5239a-kube-api-access-8rt98" (OuterVolumeSpecName: "kube-api-access-8rt98") pod "66594344-3d62-414a-9e89-3ab843c5239a" (UID: "66594344-3d62-414a-9e89-3ab843c5239a"). InnerVolumeSpecName "kube-api-access-8rt98". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.285453 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-kube-api-access-nnpvm" (OuterVolumeSpecName: "kube-api-access-nnpvm") pod "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" (UID: "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5"). InnerVolumeSpecName "kube-api-access-nnpvm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.285515 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" (UID: "29836d6d-6ed4-4ed6-9de6-303f7bf98ab5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.305580 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fd9df75-65d5-476e-9e82-da35f1116996" (UID: "9fd9df75-65d5-476e-9e82-da35f1116996"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.323080 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72bdf6fb-83a5-4dcb-996c-0bee47903667" (UID: "72bdf6fb-83a5-4dcb-996c-0bee47903667"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.342897 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0b63bb8-f8c1-4591-affd-65e1d6cb3393" (UID: "a0b63bb8-f8c1-4591-affd-65e1d6cb3393"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375753 5110 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375791 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375804 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375815 5110 reconciler_common.go:299] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375827 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375837 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375847 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9gr2k\" (UniqueName: \"kubernetes.io/projected/9fd9df75-65d5-476e-9e82-da35f1116996-kube-api-access-9gr2k\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375858 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72bdf6fb-83a5-4dcb-996c-0bee47903667-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375870 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-czp6q\" (UniqueName: \"kubernetes.io/projected/72bdf6fb-83a5-4dcb-996c-0bee47903667-kube-api-access-czp6q\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375879 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fd9df75-65d5-476e-9e82-da35f1116996-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375889 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nnpvm\" (UniqueName: \"kubernetes.io/projected/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-kube-api-access-nnpvm\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375899 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q4hxl\" (UniqueName: \"kubernetes.io/projected/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-kube-api-access-q4hxl\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375910 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8rt98\" (UniqueName: \"kubernetes.io/projected/66594344-3d62-414a-9e89-3ab843c5239a-kube-api-access-8rt98\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375922 5110 reconciler_common.go:299] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.375932 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b63bb8-f8c1-4591-affd-65e1d6cb3393-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.404582 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66594344-3d62-414a-9e89-3ab843c5239a" (UID: "66594344-3d62-414a-9e89-3ab843c5239a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.476893 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66594344-3d62-414a-9e89-3ab843c5239a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.572359 5110 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.572516 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-9m47z"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.850673 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" event={"ID":"53db1455-0000-4bb3-9f81-0e079a38a5ec","Type":"ContainerStarted","Data":"fc3406c253db0d6a16c942c82ec73c08c62f64eafdbcbb2911e63f99e6540470"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.850737 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" event={"ID":"53db1455-0000-4bb3-9f81-0e079a38a5ec","Type":"ContainerStarted","Data":"c73208cd81b0e71e417937081322e7d8a7b75960a7129b3ba36d6a886e7d4270"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.850872 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.851999 5110 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-9m47z container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.69:8080/healthz\": dial tcp 10.217.0.69:8080: connect: connection refused" start-of-body= Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.852091 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" podUID="53db1455-0000-4bb3-9f81-0e079a38a5ec" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.69:8080/healthz\": dial tcp 10.217.0.69:8080: connect: connection refused" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.853806 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.853876 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-95rbr" event={"ID":"29836d6d-6ed4-4ed6-9de6-303f7bf98ab5","Type":"ContainerDied","Data":"68f777267fb4ec1644dd1ad144840b6905cfe1a4c3e41ce8c8d77b29e0b4bb97"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.853917 5110 scope.go:117] "RemoveContainer" containerID="d5ce783b9323ff971b5b6c0469ad3e4208d7689e5474a70f8f41e7093fdf50d8" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.858368 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh5rg" event={"ID":"9fd9df75-65d5-476e-9e82-da35f1116996","Type":"ContainerDied","Data":"ee503274009754b42e63290144f3ac7ab5adc668a60a9feeae54ec857a788f04"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.858482 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh5rg" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.863844 5110 generic.go:358] "Generic (PLEG): container finished" podID="66594344-3d62-414a-9e89-3ab843c5239a" containerID="3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9" exitCode=0 Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.863933 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerDied","Data":"3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.863958 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctgs" event={"ID":"66594344-3d62-414a-9e89-3ab843c5239a","Type":"ContainerDied","Data":"7dd9f9bea04bda9234d2c6ee3bcc6614399452cbdd04c713668449b94d711a0a"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.864035 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctgs" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.866783 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czltc" event={"ID":"a0b63bb8-f8c1-4591-affd-65e1d6cb3393","Type":"ContainerDied","Data":"13ddd43aea779eaf69faa20c0ea571ded8a9fa997b9cd5977b2a945417f7cdf5"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.866898 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-czltc" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.869274 5110 scope.go:117] "RemoveContainer" containerID="2808e465d7090bf5dc3dabbb865ad0a0bc592a6ff26bb912931ee02f7f4af634" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.874202 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldd2" event={"ID":"72bdf6fb-83a5-4dcb-996c-0bee47903667","Type":"ContainerDied","Data":"8bc77024ac1a104eadb739838a88ed32b8d771568727e3c40767b594bdb41554"} Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.874284 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldd2" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.886882 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" podStartSLOduration=1.886868894 podStartE2EDuration="1.886868894s" podCreationTimestamp="2026-03-17 18:53:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:53:11.875155761 +0000 UTC m=+335.907767293" watchObservedRunningTime="2026-03-17 18:53:11.886868894 +0000 UTC m=+335.919480416" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.906978 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-95rbr"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.910823 5110 scope.go:117] "RemoveContainer" containerID="0ecc64867c73b49e8d7fd6b12ce965282df07aaa3c03a503afb080d528c217b2" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.917227 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-95rbr"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.927207 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh5rg"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.932227 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh5rg"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.944996 5110 scope.go:117] "RemoveContainer" containerID="79088d6b5a1462e75a1de3a8d7242f8f4ba59110b5dbda757fba96e3f8771a70" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.964518 5110 scope.go:117] "RemoveContainer" containerID="3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.970191 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nldd2"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.979458 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nldd2"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.992367 5110 scope.go:117] "RemoveContainer" containerID="fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c" Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.992465 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-czltc"] Mar 17 18:53:11 crc kubenswrapper[5110]: I0317 18:53:11.997444 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-czltc"] Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.000968 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jctgs"] Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.003771 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jctgs"] Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.012407 5110 scope.go:117] "RemoveContainer" containerID="b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.029365 5110 scope.go:117] "RemoveContainer" containerID="3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9" Mar 17 18:53:12 crc kubenswrapper[5110]: E0317 18:53:12.029767 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9\": container with ID starting with 3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9 not found: ID does not exist" containerID="3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.029809 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9"} err="failed to get container status \"3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9\": rpc error: code = NotFound desc = could not find container \"3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9\": container with ID starting with 3db009a93783be46b7c9177c4428924fa1ac098d6253e3ea914f19cbe1ea16f9 not found: ID does not exist" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.029833 5110 scope.go:117] "RemoveContainer" containerID="fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c" Mar 17 18:53:12 crc kubenswrapper[5110]: E0317 18:53:12.030138 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c\": container with ID starting with fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c not found: ID does not exist" containerID="fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.030185 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c"} err="failed to get container status \"fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c\": rpc error: code = NotFound desc = could not find container \"fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c\": container with ID starting with fbac867abdbe7bc471763e98cc56837a3d6ca948cbe1f2e6c23026042695bb0c not found: ID does not exist" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.030210 5110 scope.go:117] "RemoveContainer" containerID="b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8" Mar 17 18:53:12 crc kubenswrapper[5110]: E0317 18:53:12.030508 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8\": container with ID starting with b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8 not found: ID does not exist" containerID="b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.030537 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8"} err="failed to get container status \"b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8\": rpc error: code = NotFound desc = could not find container \"b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8\": container with ID starting with b11d4362d301bb8855c474bbd06752d1a73544fefb96b15f1aaf4b0c8e57d7e8 not found: ID does not exist" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.030556 5110 scope.go:117] "RemoveContainer" containerID="052c29ca4712916b5105c422bb78da12e80470450fd5d29aeda39dd4bef5d3e6" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.042979 5110 scope.go:117] "RemoveContainer" containerID="6eed216aebd0ab34f7b96032e984f7c4b7620698696d48a20bd480014b415183" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.055041 5110 scope.go:117] "RemoveContainer" containerID="ae938c8d26b6fc6a3dc1dc7c2ca94d7be64cc69517a79ec75da05d44542ea2c7" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.068328 5110 scope.go:117] "RemoveContainer" containerID="5ddd0af732769f3a01a5f513e450a9a6ed4a6589e2c0cb2a9bfb40f8d8b14e56" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.087445 5110 scope.go:117] "RemoveContainer" containerID="b7664d5b2c2744a1bd9fe29f53fdbd2c44c70ae044074eec0151f5fedd1b1447" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.109014 5110 scope.go:117] "RemoveContainer" containerID="be05241bad40425d9a516f897f859b0220916a46d99d78693d71dd29ba1aa337" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.562216 5110 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.757762 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" path="/var/lib/kubelet/pods/29836d6d-6ed4-4ed6-9de6-303f7bf98ab5/volumes" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.758391 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66594344-3d62-414a-9e89-3ab843c5239a" path="/var/lib/kubelet/pods/66594344-3d62-414a-9e89-3ab843c5239a/volumes" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.758913 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" path="/var/lib/kubelet/pods/72bdf6fb-83a5-4dcb-996c-0bee47903667/volumes" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.759500 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" path="/var/lib/kubelet/pods/9fd9df75-65d5-476e-9e82-da35f1116996/volumes" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.760017 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" path="/var/lib/kubelet/pods/a0b63bb8-f8c1-4591-affd-65e1d6cb3393/volumes" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.874260 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-krsp5"] Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.875876 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.876017 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.877965 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878381 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878473 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878565 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878644 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878711 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878796 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878876 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.878958 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879043 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879149 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879227 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879301 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879408 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879490 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879654 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="extract-utilities" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879741 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879812 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879886 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.879955 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880043 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880149 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880234 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880313 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="extract-content" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880393 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880470 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880706 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880794 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="66594344-3d62-414a-9e89-3ab843c5239a" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880876 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="29836d6d-6ed4-4ed6-9de6-303f7bf98ab5" containerName="marketplace-operator" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.880957 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="9fd9df75-65d5-476e-9e82-da35f1116996" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.881032 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="a0b63bb8-f8c1-4591-affd-65e1d6cb3393" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.881144 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="72bdf6fb-83a5-4dcb-996c-0bee47903667" containerName="registry-server" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.891985 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krsp5"] Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.892231 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.896332 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Mar 17 18:53:12 crc kubenswrapper[5110]: I0317 18:53:12.909466 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-9m47z" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.003350 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-utilities\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.003498 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-catalog-content\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.003654 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-kube-api-access-fkrrs\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.078171 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pfbcf"] Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.093253 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pfbcf"] Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.093385 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.095683 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.104589 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-kube-api-access-fkrrs\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.104657 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-utilities\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.104707 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-catalog-content\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.105215 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-catalog-content\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.105665 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-utilities\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.130048 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-kube-api-access-fkrrs\") pod \"redhat-marketplace-krsp5\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.206302 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-catalog-content\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.206380 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-utilities\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.206398 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mj26\" (UniqueName: \"kubernetes.io/projected/0861862c-403b-4786-bc1d-0c9d9fab15b5-kube-api-access-9mj26\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.211121 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.308552 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-utilities\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.308746 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9mj26\" (UniqueName: \"kubernetes.io/projected/0861862c-403b-4786-bc1d-0c9d9fab15b5-kube-api-access-9mj26\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.308845 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-catalog-content\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.309431 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-catalog-content\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.309712 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-utilities\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.334933 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mj26\" (UniqueName: \"kubernetes.io/projected/0861862c-403b-4786-bc1d-0c9d9fab15b5-kube-api-access-9mj26\") pod \"certified-operators-pfbcf\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.414664 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.642887 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krsp5"] Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.796613 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pfbcf"] Mar 17 18:53:13 crc kubenswrapper[5110]: W0317 18:53:13.820471 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0861862c_403b_4786_bc1d_0c9d9fab15b5.slice/crio-c8e3054df075c38686996efbc315dccdb925300191e33f692944d70821661f13 WatchSource:0}: Error finding container c8e3054df075c38686996efbc315dccdb925300191e33f692944d70821661f13: Status 404 returned error can't find the container with id c8e3054df075c38686996efbc315dccdb925300191e33f692944d70821661f13 Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.894351 5110 generic.go:358] "Generic (PLEG): container finished" podID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerID="c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea" exitCode=0 Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.894472 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krsp5" event={"ID":"fc56a95b-a15c-4661-8247-0ec3d89ac2b6","Type":"ContainerDied","Data":"c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea"} Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.894497 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krsp5" event={"ID":"fc56a95b-a15c-4661-8247-0ec3d89ac2b6","Type":"ContainerStarted","Data":"d9c88f9e465b8429f90d6a0de781e7a506d2786144da9d47a162df6017f81476"} Mar 17 18:53:13 crc kubenswrapper[5110]: I0317 18:53:13.898729 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerStarted","Data":"c8e3054df075c38686996efbc315dccdb925300191e33f692944d70821661f13"} Mar 17 18:53:14 crc kubenswrapper[5110]: I0317 18:53:14.906148 5110 generic.go:358] "Generic (PLEG): container finished" podID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerID="b4f7d9e5953553869617712e07b22a9b9b8c49e19ccf09d006620c94336b54d1" exitCode=0 Mar 17 18:53:14 crc kubenswrapper[5110]: I0317 18:53:14.906246 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerDied","Data":"b4f7d9e5953553869617712e07b22a9b9b8c49e19ccf09d006620c94336b54d1"} Mar 17 18:53:14 crc kubenswrapper[5110]: I0317 18:53:14.911029 5110 generic.go:358] "Generic (PLEG): container finished" podID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerID="f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402" exitCode=0 Mar 17 18:53:14 crc kubenswrapper[5110]: I0317 18:53:14.911124 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krsp5" event={"ID":"fc56a95b-a15c-4661-8247-0ec3d89ac2b6","Type":"ContainerDied","Data":"f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402"} Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.271641 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hvkkc"] Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.279914 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.282574 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.299828 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvkkc"] Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.338147 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293dea77-4f08-44cc-b93c-7eb90f01c833-utilities\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.338208 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt8n2\" (UniqueName: \"kubernetes.io/projected/293dea77-4f08-44cc-b93c-7eb90f01c833-kube-api-access-bt8n2\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.338538 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293dea77-4f08-44cc-b93c-7eb90f01c833-catalog-content\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.439628 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293dea77-4f08-44cc-b93c-7eb90f01c833-catalog-content\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.439679 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293dea77-4f08-44cc-b93c-7eb90f01c833-utilities\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.439706 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bt8n2\" (UniqueName: \"kubernetes.io/projected/293dea77-4f08-44cc-b93c-7eb90f01c833-kube-api-access-bt8n2\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.440277 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293dea77-4f08-44cc-b93c-7eb90f01c833-catalog-content\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.440456 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293dea77-4f08-44cc-b93c-7eb90f01c833-utilities\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.463333 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt8n2\" (UniqueName: \"kubernetes.io/projected/293dea77-4f08-44cc-b93c-7eb90f01c833-kube-api-access-bt8n2\") pod \"redhat-operators-hvkkc\" (UID: \"293dea77-4f08-44cc-b93c-7eb90f01c833\") " pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.495130 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t72j2"] Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.501348 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.503740 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.505927 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t72j2"] Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.540906 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-utilities\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.540965 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llnxd\" (UniqueName: \"kubernetes.io/projected/deba3474-2676-42bb-8a3a-5d47a06caf16-kube-api-access-llnxd\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.541090 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-catalog-content\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.604992 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.642261 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-catalog-content\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.642738 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-utilities\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.642802 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-llnxd\" (UniqueName: \"kubernetes.io/projected/deba3474-2676-42bb-8a3a-5d47a06caf16-kube-api-access-llnxd\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.643626 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-utilities\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.643763 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-catalog-content\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.669090 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-llnxd\" (UniqueName: \"kubernetes.io/projected/deba3474-2676-42bb-8a3a-5d47a06caf16-kube-api-access-llnxd\") pod \"community-operators-t72j2\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.820408 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.924071 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerStarted","Data":"5624f47da017bf23fc43ad49f2c445de17871aa9cfc2ad2266c690145d0cd46f"} Mar 17 18:53:15 crc kubenswrapper[5110]: I0317 18:53:15.928237 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krsp5" event={"ID":"fc56a95b-a15c-4661-8247-0ec3d89ac2b6","Type":"ContainerStarted","Data":"91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67"} Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.019345 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-krsp5" podStartSLOduration=3.382888492 podStartE2EDuration="4.019322029s" podCreationTimestamp="2026-03-17 18:53:12 +0000 UTC" firstStartedPulling="2026-03-17 18:53:13.894970167 +0000 UTC m=+337.927581689" lastFinishedPulling="2026-03-17 18:53:14.531403704 +0000 UTC m=+338.564015226" observedRunningTime="2026-03-17 18:53:15.967427877 +0000 UTC m=+340.000039409" watchObservedRunningTime="2026-03-17 18:53:16.019322029 +0000 UTC m=+340.051933561" Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.022929 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvkkc"] Mar 17 18:53:16 crc kubenswrapper[5110]: W0317 18:53:16.038664 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod293dea77_4f08_44cc_b93c_7eb90f01c833.slice/crio-0ce44d85878c0ddd85e631aceb0580a8f3386a1ea0ffdd6a2f62e123eb75c220 WatchSource:0}: Error finding container 0ce44d85878c0ddd85e631aceb0580a8f3386a1ea0ffdd6a2f62e123eb75c220: Status 404 returned error can't find the container with id 0ce44d85878c0ddd85e631aceb0580a8f3386a1ea0ffdd6a2f62e123eb75c220 Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.205777 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t72j2"] Mar 17 18:53:16 crc kubenswrapper[5110]: W0317 18:53:16.220687 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeba3474_2676_42bb_8a3a_5d47a06caf16.slice/crio-74e27a1bf21c1a283e13031e84124a6632570f8192afdf2bc205f5f90cc42529 WatchSource:0}: Error finding container 74e27a1bf21c1a283e13031e84124a6632570f8192afdf2bc205f5f90cc42529: Status 404 returned error can't find the container with id 74e27a1bf21c1a283e13031e84124a6632570f8192afdf2bc205f5f90cc42529 Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.936410 5110 generic.go:358] "Generic (PLEG): container finished" podID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerID="d4c0a61b0a353897a51ba364aa8442cca024eb2c6e0f6711eb50652942ec4e27" exitCode=0 Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.936456 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerDied","Data":"d4c0a61b0a353897a51ba364aa8442cca024eb2c6e0f6711eb50652942ec4e27"} Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.936494 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerStarted","Data":"74e27a1bf21c1a283e13031e84124a6632570f8192afdf2bc205f5f90cc42529"} Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.941930 5110 generic.go:358] "Generic (PLEG): container finished" podID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerID="5624f47da017bf23fc43ad49f2c445de17871aa9cfc2ad2266c690145d0cd46f" exitCode=0 Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.942113 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerDied","Data":"5624f47da017bf23fc43ad49f2c445de17871aa9cfc2ad2266c690145d0cd46f"} Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.944375 5110 generic.go:358] "Generic (PLEG): container finished" podID="293dea77-4f08-44cc-b93c-7eb90f01c833" containerID="c7401bedcbbaa25957695a9cab27eee5e550b183b50f70d7b72695937025bdff" exitCode=0 Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.945291 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvkkc" event={"ID":"293dea77-4f08-44cc-b93c-7eb90f01c833","Type":"ContainerDied","Data":"c7401bedcbbaa25957695a9cab27eee5e550b183b50f70d7b72695937025bdff"} Mar 17 18:53:16 crc kubenswrapper[5110]: I0317 18:53:16.945340 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvkkc" event={"ID":"293dea77-4f08-44cc-b93c-7eb90f01c833","Type":"ContainerStarted","Data":"0ce44d85878c0ddd85e631aceb0580a8f3386a1ea0ffdd6a2f62e123eb75c220"} Mar 17 18:53:17 crc kubenswrapper[5110]: I0317 18:53:17.952678 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerStarted","Data":"39128367db3304d0be722eac652f6b3db3e7144568325a3ca0cef638f0905454"} Mar 17 18:53:17 crc kubenswrapper[5110]: I0317 18:53:17.956102 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerStarted","Data":"42e7b9b2ea6d754a5623e9e547bcd92a370638378f4c1bffef9e57065f4f924e"} Mar 17 18:53:17 crc kubenswrapper[5110]: I0317 18:53:17.957993 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvkkc" event={"ID":"293dea77-4f08-44cc-b93c-7eb90f01c833","Type":"ContainerStarted","Data":"32fda6a19cb1ec60a76473c0026a135bab7cdb6130e48a59f89d55d093a52f20"} Mar 17 18:53:17 crc kubenswrapper[5110]: I0317 18:53:17.997127 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pfbcf" podStartSLOduration=4.265784463 podStartE2EDuration="4.997107662s" podCreationTimestamp="2026-03-17 18:53:13 +0000 UTC" firstStartedPulling="2026-03-17 18:53:14.907810343 +0000 UTC m=+338.940421895" lastFinishedPulling="2026-03-17 18:53:15.639133572 +0000 UTC m=+339.671745094" observedRunningTime="2026-03-17 18:53:17.993526201 +0000 UTC m=+342.026137763" watchObservedRunningTime="2026-03-17 18:53:17.997107662 +0000 UTC m=+342.029719184" Mar 17 18:53:18 crc kubenswrapper[5110]: I0317 18:53:18.967739 5110 generic.go:358] "Generic (PLEG): container finished" podID="293dea77-4f08-44cc-b93c-7eb90f01c833" containerID="32fda6a19cb1ec60a76473c0026a135bab7cdb6130e48a59f89d55d093a52f20" exitCode=0 Mar 17 18:53:18 crc kubenswrapper[5110]: I0317 18:53:18.968204 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvkkc" event={"ID":"293dea77-4f08-44cc-b93c-7eb90f01c833","Type":"ContainerDied","Data":"32fda6a19cb1ec60a76473c0026a135bab7cdb6130e48a59f89d55d093a52f20"} Mar 17 18:53:18 crc kubenswrapper[5110]: I0317 18:53:18.974721 5110 generic.go:358] "Generic (PLEG): container finished" podID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerID="39128367db3304d0be722eac652f6b3db3e7144568325a3ca0cef638f0905454" exitCode=0 Mar 17 18:53:18 crc kubenswrapper[5110]: I0317 18:53:18.975117 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerDied","Data":"39128367db3304d0be722eac652f6b3db3e7144568325a3ca0cef638f0905454"} Mar 17 18:53:19 crc kubenswrapper[5110]: I0317 18:53:19.984309 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerStarted","Data":"7c9d9e3b1f7c677c263a5bff83512434f4bfd20e3ca5de0f68576f2c291fd995"} Mar 17 18:53:19 crc kubenswrapper[5110]: I0317 18:53:19.988454 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvkkc" event={"ID":"293dea77-4f08-44cc-b93c-7eb90f01c833","Type":"ContainerStarted","Data":"23dbdeda7b0df0994ac195f51bc82f48e7984c7ab8ea64e4e3e1d0ebc9656b51"} Mar 17 18:53:20 crc kubenswrapper[5110]: I0317 18:53:20.033887 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hvkkc" podStartSLOduration=4.386503944 podStartE2EDuration="5.03387117s" podCreationTimestamp="2026-03-17 18:53:15 +0000 UTC" firstStartedPulling="2026-03-17 18:53:16.946246438 +0000 UTC m=+340.978857970" lastFinishedPulling="2026-03-17 18:53:17.593613674 +0000 UTC m=+341.626225196" observedRunningTime="2026-03-17 18:53:20.030854294 +0000 UTC m=+344.063465826" watchObservedRunningTime="2026-03-17 18:53:20.03387117 +0000 UTC m=+344.066482702" Mar 17 18:53:20 crc kubenswrapper[5110]: I0317 18:53:20.035281 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t72j2" podStartSLOduration=4.453133954 podStartE2EDuration="5.035271949s" podCreationTimestamp="2026-03-17 18:53:15 +0000 UTC" firstStartedPulling="2026-03-17 18:53:16.940146945 +0000 UTC m=+340.972758477" lastFinishedPulling="2026-03-17 18:53:17.52228495 +0000 UTC m=+341.554896472" observedRunningTime="2026-03-17 18:53:20.007983935 +0000 UTC m=+344.040595467" watchObservedRunningTime="2026-03-17 18:53:20.035271949 +0000 UTC m=+344.067883481" Mar 17 18:53:23 crc kubenswrapper[5110]: I0317 18:53:23.211508 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:23 crc kubenswrapper[5110]: I0317 18:53:23.211818 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:23 crc kubenswrapper[5110]: I0317 18:53:23.293603 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:23 crc kubenswrapper[5110]: I0317 18:53:23.415406 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:23 crc kubenswrapper[5110]: I0317 18:53:23.415449 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:23 crc kubenswrapper[5110]: I0317 18:53:23.446070 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:24 crc kubenswrapper[5110]: I0317 18:53:24.071688 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 18:53:24 crc kubenswrapper[5110]: I0317 18:53:24.081759 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:53:25 crc kubenswrapper[5110]: I0317 18:53:25.606326 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:25 crc kubenswrapper[5110]: I0317 18:53:25.606410 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:25 crc kubenswrapper[5110]: I0317 18:53:25.821479 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:25 crc kubenswrapper[5110]: I0317 18:53:25.821829 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:25 crc kubenswrapper[5110]: I0317 18:53:25.858587 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:26 crc kubenswrapper[5110]: I0317 18:53:26.063597 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t72j2" Mar 17 18:53:26 crc kubenswrapper[5110]: I0317 18:53:26.662908 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hvkkc" podUID="293dea77-4f08-44cc-b93c-7eb90f01c833" containerName="registry-server" probeResult="failure" output=< Mar 17 18:53:26 crc kubenswrapper[5110]: timeout: failed to connect service ":50051" within 1s Mar 17 18:53:26 crc kubenswrapper[5110]: > Mar 17 18:53:35 crc kubenswrapper[5110]: I0317 18:53:35.643568 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:53:35 crc kubenswrapper[5110]: I0317 18:53:35.689247 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hvkkc" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.138762 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562894-qrq5w"] Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.143587 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.147374 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562894-qrq5w"] Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.147878 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.147892 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.148233 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.238979 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpkq8\" (UniqueName: \"kubernetes.io/projected/1dc544ed-7904-47b8-9980-10d7015526ac-kube-api-access-wpkq8\") pod \"auto-csr-approver-29562894-qrq5w\" (UID: \"1dc544ed-7904-47b8-9980-10d7015526ac\") " pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.340388 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wpkq8\" (UniqueName: \"kubernetes.io/projected/1dc544ed-7904-47b8-9980-10d7015526ac-kube-api-access-wpkq8\") pod \"auto-csr-approver-29562894-qrq5w\" (UID: \"1dc544ed-7904-47b8-9980-10d7015526ac\") " pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.365676 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpkq8\" (UniqueName: \"kubernetes.io/projected/1dc544ed-7904-47b8-9980-10d7015526ac-kube-api-access-wpkq8\") pod \"auto-csr-approver-29562894-qrq5w\" (UID: \"1dc544ed-7904-47b8-9980-10d7015526ac\") " pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.471349 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:00 crc kubenswrapper[5110]: I0317 18:54:00.711428 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562894-qrq5w"] Mar 17 18:54:01 crc kubenswrapper[5110]: I0317 18:54:01.223840 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" event={"ID":"1dc544ed-7904-47b8-9980-10d7015526ac","Type":"ContainerStarted","Data":"7fe26a27421618e704bb267eb3584c09ab92c0fead871c54c5dfb3bdec9ebb58"} Mar 17 18:54:02 crc kubenswrapper[5110]: I0317 18:54:02.231293 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" event={"ID":"1dc544ed-7904-47b8-9980-10d7015526ac","Type":"ContainerStarted","Data":"3d5a1402655aaf02e57e9eecb26b8ba788c0471bd91911a79c43849c55f8d8a5"} Mar 17 18:54:02 crc kubenswrapper[5110]: I0317 18:54:02.247786 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" podStartSLOduration=1.093086896 podStartE2EDuration="2.247769927s" podCreationTimestamp="2026-03-17 18:54:00 +0000 UTC" firstStartedPulling="2026-03-17 18:54:00.7372645 +0000 UTC m=+384.769876032" lastFinishedPulling="2026-03-17 18:54:01.891947541 +0000 UTC m=+385.924559063" observedRunningTime="2026-03-17 18:54:02.244986118 +0000 UTC m=+386.277597660" watchObservedRunningTime="2026-03-17 18:54:02.247769927 +0000 UTC m=+386.280381449" Mar 17 18:54:03 crc kubenswrapper[5110]: I0317 18:54:03.237130 5110 generic.go:358] "Generic (PLEG): container finished" podID="1dc544ed-7904-47b8-9980-10d7015526ac" containerID="3d5a1402655aaf02e57e9eecb26b8ba788c0471bd91911a79c43849c55f8d8a5" exitCode=0 Mar 17 18:54:03 crc kubenswrapper[5110]: I0317 18:54:03.237260 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" event={"ID":"1dc544ed-7904-47b8-9980-10d7015526ac","Type":"ContainerDied","Data":"3d5a1402655aaf02e57e9eecb26b8ba788c0471bd91911a79c43849c55f8d8a5"} Mar 17 18:54:04 crc kubenswrapper[5110]: I0317 18:54:04.546590 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:04 crc kubenswrapper[5110]: I0317 18:54:04.594934 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpkq8\" (UniqueName: \"kubernetes.io/projected/1dc544ed-7904-47b8-9980-10d7015526ac-kube-api-access-wpkq8\") pod \"1dc544ed-7904-47b8-9980-10d7015526ac\" (UID: \"1dc544ed-7904-47b8-9980-10d7015526ac\") " Mar 17 18:54:04 crc kubenswrapper[5110]: I0317 18:54:04.600462 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc544ed-7904-47b8-9980-10d7015526ac-kube-api-access-wpkq8" (OuterVolumeSpecName: "kube-api-access-wpkq8") pod "1dc544ed-7904-47b8-9980-10d7015526ac" (UID: "1dc544ed-7904-47b8-9980-10d7015526ac"). InnerVolumeSpecName "kube-api-access-wpkq8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:54:04 crc kubenswrapper[5110]: I0317 18:54:04.696332 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wpkq8\" (UniqueName: \"kubernetes.io/projected/1dc544ed-7904-47b8-9980-10d7015526ac-kube-api-access-wpkq8\") on node \"crc\" DevicePath \"\"" Mar 17 18:54:05 crc kubenswrapper[5110]: I0317 18:54:05.252356 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" event={"ID":"1dc544ed-7904-47b8-9980-10d7015526ac","Type":"ContainerDied","Data":"7fe26a27421618e704bb267eb3584c09ab92c0fead871c54c5dfb3bdec9ebb58"} Mar 17 18:54:05 crc kubenswrapper[5110]: I0317 18:54:05.252413 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fe26a27421618e704bb267eb3584c09ab92c0fead871c54c5dfb3bdec9ebb58" Mar 17 18:54:05 crc kubenswrapper[5110]: I0317 18:54:05.252485 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562894-qrq5w" Mar 17 18:54:12 crc kubenswrapper[5110]: I0317 18:54:12.133637 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:54:12 crc kubenswrapper[5110]: I0317 18:54:12.134395 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:54:42 crc kubenswrapper[5110]: I0317 18:54:42.132972 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:54:42 crc kubenswrapper[5110]: I0317 18:54:42.134042 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.133168 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.133799 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.133867 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.134800 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ebe54d3668f3ba6d72d2f6801190bcc6cdfdce7f9ffcdbc79df03266e49b4ac1"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.134911 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://ebe54d3668f3ba6d72d2f6801190bcc6cdfdce7f9ffcdbc79df03266e49b4ac1" gracePeriod=600 Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.701434 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="ebe54d3668f3ba6d72d2f6801190bcc6cdfdce7f9ffcdbc79df03266e49b4ac1" exitCode=0 Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.701526 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"ebe54d3668f3ba6d72d2f6801190bcc6cdfdce7f9ffcdbc79df03266e49b4ac1"} Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.701853 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"6c003f768dfcdd99f40db6c78311c785574f0d09dbaf1d45833991c2b96d1903"} Mar 17 18:55:12 crc kubenswrapper[5110]: I0317 18:55:12.701875 5110 scope.go:117] "RemoveContainer" containerID="0075007a0a6d1dcc16449135cf348101cd8615648447feb043c5978c7d129075" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.141039 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562896-h6cq4"] Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.142346 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1dc544ed-7904-47b8-9980-10d7015526ac" containerName="oc" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.142365 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc544ed-7904-47b8-9980-10d7015526ac" containerName="oc" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.142495 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="1dc544ed-7904-47b8-9980-10d7015526ac" containerName="oc" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.158198 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562896-h6cq4"] Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.158326 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.162133 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.163226 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.165766 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.220237 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2zt8\" (UniqueName: \"kubernetes.io/projected/8093480b-748b-460c-93cf-ed80e5a3f4ec-kube-api-access-g2zt8\") pod \"auto-csr-approver-29562896-h6cq4\" (UID: \"8093480b-748b-460c-93cf-ed80e5a3f4ec\") " pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.321362 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g2zt8\" (UniqueName: \"kubernetes.io/projected/8093480b-748b-460c-93cf-ed80e5a3f4ec-kube-api-access-g2zt8\") pod \"auto-csr-approver-29562896-h6cq4\" (UID: \"8093480b-748b-460c-93cf-ed80e5a3f4ec\") " pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.345981 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2zt8\" (UniqueName: \"kubernetes.io/projected/8093480b-748b-460c-93cf-ed80e5a3f4ec-kube-api-access-g2zt8\") pod \"auto-csr-approver-29562896-h6cq4\" (UID: \"8093480b-748b-460c-93cf-ed80e5a3f4ec\") " pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.511099 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:00 crc kubenswrapper[5110]: I0317 18:56:00.730151 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562896-h6cq4"] Mar 17 18:56:01 crc kubenswrapper[5110]: I0317 18:56:01.009855 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" event={"ID":"8093480b-748b-460c-93cf-ed80e5a3f4ec","Type":"ContainerStarted","Data":"cf4922ca461fdc37a205cd679020e59ba8bcc29adffb0b83fd299b9036b8e881"} Mar 17 18:56:03 crc kubenswrapper[5110]: I0317 18:56:03.028522 5110 generic.go:358] "Generic (PLEG): container finished" podID="8093480b-748b-460c-93cf-ed80e5a3f4ec" containerID="bdf416706f8b9b802ac35fcb6bbb69fad04da677b31864ece050af098f442aa4" exitCode=0 Mar 17 18:56:03 crc kubenswrapper[5110]: I0317 18:56:03.028607 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" event={"ID":"8093480b-748b-460c-93cf-ed80e5a3f4ec","Type":"ContainerDied","Data":"bdf416706f8b9b802ac35fcb6bbb69fad04da677b31864ece050af098f442aa4"} Mar 17 18:56:04 crc kubenswrapper[5110]: I0317 18:56:04.249443 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:04 crc kubenswrapper[5110]: I0317 18:56:04.282680 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2zt8\" (UniqueName: \"kubernetes.io/projected/8093480b-748b-460c-93cf-ed80e5a3f4ec-kube-api-access-g2zt8\") pod \"8093480b-748b-460c-93cf-ed80e5a3f4ec\" (UID: \"8093480b-748b-460c-93cf-ed80e5a3f4ec\") " Mar 17 18:56:04 crc kubenswrapper[5110]: I0317 18:56:04.289996 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8093480b-748b-460c-93cf-ed80e5a3f4ec-kube-api-access-g2zt8" (OuterVolumeSpecName: "kube-api-access-g2zt8") pod "8093480b-748b-460c-93cf-ed80e5a3f4ec" (UID: "8093480b-748b-460c-93cf-ed80e5a3f4ec"). InnerVolumeSpecName "kube-api-access-g2zt8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:56:04 crc kubenswrapper[5110]: I0317 18:56:04.384185 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-g2zt8\" (UniqueName: \"kubernetes.io/projected/8093480b-748b-460c-93cf-ed80e5a3f4ec-kube-api-access-g2zt8\") on node \"crc\" DevicePath \"\"" Mar 17 18:56:05 crc kubenswrapper[5110]: I0317 18:56:05.043608 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" Mar 17 18:56:05 crc kubenswrapper[5110]: I0317 18:56:05.043700 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562896-h6cq4" event={"ID":"8093480b-748b-460c-93cf-ed80e5a3f4ec","Type":"ContainerDied","Data":"cf4922ca461fdc37a205cd679020e59ba8bcc29adffb0b83fd299b9036b8e881"} Mar 17 18:56:05 crc kubenswrapper[5110]: I0317 18:56:05.044175 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf4922ca461fdc37a205cd679020e59ba8bcc29adffb0b83fd299b9036b8e881" Mar 17 18:56:05 crc kubenswrapper[5110]: I0317 18:56:05.318481 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562890-xwcc2"] Mar 17 18:56:05 crc kubenswrapper[5110]: I0317 18:56:05.323591 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562890-xwcc2"] Mar 17 18:56:06 crc kubenswrapper[5110]: I0317 18:56:06.752812 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44f42740-7b5b-4e4f-9e9e-c4c83b6c1575" path="/var/lib/kubelet/pods/44f42740-7b5b-4e4f-9e9e-c4c83b6c1575/volumes" Mar 17 18:56:37 crc kubenswrapper[5110]: I0317 18:56:37.192671 5110 scope.go:117] "RemoveContainer" containerID="f4b4b2b33f22225cd1c4788078e50e75b65f3a5923ea97ad909adbe4ec3e98e8" Mar 17 18:57:12 crc kubenswrapper[5110]: I0317 18:57:12.132880 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:57:12 crc kubenswrapper[5110]: I0317 18:57:12.134276 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:57:37 crc kubenswrapper[5110]: I0317 18:57:37.029288 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 18:57:37 crc kubenswrapper[5110]: I0317 18:57:37.035959 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 18:57:42 crc kubenswrapper[5110]: I0317 18:57:42.132948 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:57:42 crc kubenswrapper[5110]: I0317 18:57:42.133669 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.138441 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562898-5lc6h"] Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.139422 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8093480b-748b-460c-93cf-ed80e5a3f4ec" containerName="oc" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.139439 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8093480b-748b-460c-93cf-ed80e5a3f4ec" containerName="oc" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.139560 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="8093480b-748b-460c-93cf-ed80e5a3f4ec" containerName="oc" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.150140 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.154821 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.155239 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.156174 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.158726 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562898-5lc6h"] Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.289029 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24rm2\" (UniqueName: \"kubernetes.io/projected/3e77e6a5-f972-41ad-9d00-d86d43c0176d-kube-api-access-24rm2\") pod \"auto-csr-approver-29562898-5lc6h\" (UID: \"3e77e6a5-f972-41ad-9d00-d86d43c0176d\") " pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.390328 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-24rm2\" (UniqueName: \"kubernetes.io/projected/3e77e6a5-f972-41ad-9d00-d86d43c0176d-kube-api-access-24rm2\") pod \"auto-csr-approver-29562898-5lc6h\" (UID: \"3e77e6a5-f972-41ad-9d00-d86d43c0176d\") " pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.410319 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-24rm2\" (UniqueName: \"kubernetes.io/projected/3e77e6a5-f972-41ad-9d00-d86d43c0176d-kube-api-access-24rm2\") pod \"auto-csr-approver-29562898-5lc6h\" (UID: \"3e77e6a5-f972-41ad-9d00-d86d43c0176d\") " pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.473522 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.682680 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562898-5lc6h"] Mar 17 18:58:00 crc kubenswrapper[5110]: I0317 18:58:00.787929 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" event={"ID":"3e77e6a5-f972-41ad-9d00-d86d43c0176d","Type":"ContainerStarted","Data":"3d05d919ecba0416f9bf60ef9fa1a77eb705bc0bc1898685928d4152ae8c788f"} Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.731372 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488"] Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.731891 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="kube-rbac-proxy" containerID="cri-o://1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.732338 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="ovnkube-cluster-manager" containerID="cri-o://dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.799098 5110 generic.go:358] "Generic (PLEG): container finished" podID="3e77e6a5-f972-41ad-9d00-d86d43c0176d" containerID="c28b96f67244aad2ab7aeeeafb8aa8e4547040532f58078ded1ff354716699fa" exitCode=0 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.799315 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" event={"ID":"3e77e6a5-f972-41ad-9d00-d86d43c0176d","Type":"ContainerDied","Data":"c28b96f67244aad2ab7aeeeafb8aa8e4547040532f58078ded1ff354716699fa"} Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.944936 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4jphj"] Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.945642 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="nbdb" containerID="cri-o://68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.946183 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="sbdb" containerID="cri-o://888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.946283 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.946352 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="northd" containerID="cri-o://94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.946414 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-node" containerID="cri-o://3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.946479 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-acl-logging" containerID="cri-o://9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.946844 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-controller" containerID="cri-o://9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.956897 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.986501 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x"] Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.987102 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="ovnkube-cluster-manager" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.987122 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="ovnkube-cluster-manager" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.987145 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="kube-rbac-proxy" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.987153 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="kube-rbac-proxy" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.987255 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="kube-rbac-proxy" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.987270 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerName="ovnkube-cluster-manager" Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.988814 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovnkube-controller" containerID="cri-o://0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" gracePeriod=30 Mar 17 18:58:02 crc kubenswrapper[5110]: I0317 18:58:02.994923 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.023362 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2nhm\" (UniqueName: \"kubernetes.io/projected/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-kube-api-access-d2nhm\") pod \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.023795 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-env-overrides\") pod \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.023910 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovnkube-config\") pod \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.023985 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovn-control-plane-metrics-cert\") pod \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\" (UID: \"fe3aacec-3e82-49dc-833e-4b1d848ea2c6\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.024448 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "fe3aacec-3e82-49dc-833e-4b1d848ea2c6" (UID: "fe3aacec-3e82-49dc-833e-4b1d848ea2c6"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.024877 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "fe3aacec-3e82-49dc-833e-4b1d848ea2c6" (UID: "fe3aacec-3e82-49dc-833e-4b1d848ea2c6"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.030041 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "fe3aacec-3e82-49dc-833e-4b1d848ea2c6" (UID: "fe3aacec-3e82-49dc-833e-4b1d848ea2c6"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.030040 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-kube-api-access-d2nhm" (OuterVolumeSpecName: "kube-api-access-d2nhm") pod "fe3aacec-3e82-49dc-833e-4b1d848ea2c6" (UID: "fe3aacec-3e82-49dc-833e-4b1d848ea2c6"). InnerVolumeSpecName "kube-api-access-d2nhm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158654 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c843896b-4355-482f-82a5-6cbd26e49d2b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158767 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c843896b-4355-482f-82a5-6cbd26e49d2b-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158858 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c843896b-4355-482f-82a5-6cbd26e49d2b-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158891 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhwvc\" (UniqueName: \"kubernetes.io/projected/c843896b-4355-482f-82a5-6cbd26e49d2b-kube-api-access-rhwvc\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158953 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d2nhm\" (UniqueName: \"kubernetes.io/projected/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-kube-api-access-d2nhm\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158969 5110 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158980 5110 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.158992 5110 reconciler_common.go:299] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fe3aacec-3e82-49dc-833e-4b1d848ea2c6-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.259682 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c843896b-4355-482f-82a5-6cbd26e49d2b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.259745 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c843896b-4355-482f-82a5-6cbd26e49d2b-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.259777 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c843896b-4355-482f-82a5-6cbd26e49d2b-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.260067 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rhwvc\" (UniqueName: \"kubernetes.io/projected/c843896b-4355-482f-82a5-6cbd26e49d2b-kube-api-access-rhwvc\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.260613 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c843896b-4355-482f-82a5-6cbd26e49d2b-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.260625 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c843896b-4355-482f-82a5-6cbd26e49d2b-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.263363 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c843896b-4355-482f-82a5-6cbd26e49d2b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.276516 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhwvc\" (UniqueName: \"kubernetes.io/projected/c843896b-4355-482f-82a5-6cbd26e49d2b-kube-api-access-rhwvc\") pod \"ovnkube-control-plane-97c9b6c48-tn26x\" (UID: \"c843896b-4355-482f-82a5-6cbd26e49d2b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.380801 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" Mar 17 18:58:03 crc kubenswrapper[5110]: W0317 18:58:03.410969 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc843896b_4355_482f_82a5_6cbd26e49d2b.slice/crio-8f1741399c46acdb16b52bd11f1b9b4a54cddcd3b6733f05be866e9252ca40b9 WatchSource:0}: Error finding container 8f1741399c46acdb16b52bd11f1b9b4a54cddcd3b6733f05be866e9252ca40b9: Status 404 returned error can't find the container with id 8f1741399c46acdb16b52bd11f1b9b4a54cddcd3b6733f05be866e9252ca40b9 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.633843 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4jphj_f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8/ovn-acl-logging/0.log" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.634382 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4jphj_f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8/ovn-controller/0.log" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.634854 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.686528 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-h6d87"] Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687492 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-controller" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687511 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-controller" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687522 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kubecfg-setup" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687528 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kubecfg-setup" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687538 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="nbdb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687543 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="nbdb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687552 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="sbdb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687557 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="sbdb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687571 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="northd" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687576 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="northd" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687584 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687590 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687601 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovnkube-controller" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687606 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovnkube-controller" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687622 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-node" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687628 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-node" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687640 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-acl-logging" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687646 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-acl-logging" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687741 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-acl-logging" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687751 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovn-controller" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687759 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="northd" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687766 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-node" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687773 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="nbdb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687779 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="sbdb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687786 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="ovnkube-controller" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.687793 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.694549 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765673 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovn-node-metrics-cert\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765715 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-slash\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765750 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-openvswitch\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765778 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-ovn-kubernetes\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765791 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-systemd\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765809 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b652\" (UniqueName: \"kubernetes.io/projected/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-kube-api-access-8b652\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765829 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-env-overrides\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765841 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-node-log\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765872 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-log-socket\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765890 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-script-lib\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765903 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-etc-openvswitch\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765942 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-systemd-units\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765967 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-kubelet\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.765985 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-netns\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766006 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-bin\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766021 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-config\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766042 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-var-lib-openvswitch\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766076 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766096 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-netd\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766131 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-ovn\") pod \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\" (UID: \"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8\") " Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766425 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766455 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-slash" (OuterVolumeSpecName: "host-slash") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766473 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766488 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.766607 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767725 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-log-socket" (OuterVolumeSpecName: "log-socket") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767774 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-node-log" (OuterVolumeSpecName: "node-log") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767845 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767874 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767906 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767909 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767927 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767914 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.767951 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.768165 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.768192 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.768219 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.771221 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-kube-api-access-8b652" (OuterVolumeSpecName: "kube-api-access-8b652") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "kube-api-access-8b652". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.771680 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.778516 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" (UID: "f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.807200 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.807293 5110 generic.go:358] "Generic (PLEG): container finished" podID="7346b312-05b5-4475-8d54-82709b69e7ed" containerID="707281b7717436044e263464eec7435119e33d38877132718dd38f6a8a0acdbb" exitCode=2 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.807455 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6kh2q" event={"ID":"7346b312-05b5-4475-8d54-82709b69e7ed","Type":"ContainerDied","Data":"707281b7717436044e263464eec7435119e33d38877132718dd38f6a8a0acdbb"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.808680 5110 scope.go:117] "RemoveContainer" containerID="707281b7717436044e263464eec7435119e33d38877132718dd38f6a8a0acdbb" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.813840 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4jphj_f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8/ovn-acl-logging/0.log" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816401 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4jphj_f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8/ovn-controller/0.log" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816911 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816941 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816971 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816979 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816986 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.816997 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817005 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" exitCode=143 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817013 5110 generic.go:358] "Generic (PLEG): container finished" podID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" exitCode=143 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817095 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817158 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817176 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817191 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817194 5110 scope.go:117] "RemoveContainer" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817207 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817220 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817236 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817250 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817257 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817240 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817267 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817373 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817385 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817394 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817400 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817408 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817414 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817420 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817427 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817433 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817453 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817468 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817476 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817483 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817489 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817496 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817501 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817509 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817515 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817521 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817530 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4jphj" event={"ID":"f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8","Type":"ContainerDied","Data":"d9fdf49e963657295a75c522c5fb9bc94198871a002c11e3280ef9eb82012cdc"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817540 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817547 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817554 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817561 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817567 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817573 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817579 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817585 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.817591 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824376 5110 generic.go:358] "Generic (PLEG): container finished" podID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerID="dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824416 5110 generic.go:358] "Generic (PLEG): container finished" podID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" containerID="1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862" exitCode=0 Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824564 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824570 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" event={"ID":"fe3aacec-3e82-49dc-833e-4b1d848ea2c6","Type":"ContainerDied","Data":"dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824600 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824611 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824626 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" event={"ID":"fe3aacec-3e82-49dc-833e-4b1d848ea2c6","Type":"ContainerDied","Data":"1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824636 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824643 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824651 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488" event={"ID":"fe3aacec-3e82-49dc-833e-4b1d848ea2c6","Type":"ContainerDied","Data":"98f18290f31e5aaa7954e356a85a85b8e221a7703ec1956532614cb7401153aa"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824668 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.824674 5110 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.831439 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" event={"ID":"c843896b-4355-482f-82a5-6cbd26e49d2b","Type":"ContainerStarted","Data":"1c90ee9420a2f5198cdffdcc520eb931ef1c5b6e467746d302f6fdf37463e353"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.831493 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" event={"ID":"c843896b-4355-482f-82a5-6cbd26e49d2b","Type":"ContainerStarted","Data":"aa4cff79d6c559d5e8e6bac56a917e387794f70f0985427182b2d05131ad93a8"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.831505 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" event={"ID":"c843896b-4355-482f-82a5-6cbd26e49d2b","Type":"ContainerStarted","Data":"8f1741399c46acdb16b52bd11f1b9b4a54cddcd3b6733f05be866e9252ca40b9"} Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.840159 5110 scope.go:117] "RemoveContainer" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.858429 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-tn26x" podStartSLOduration=1.858411847 podStartE2EDuration="1.858411847s" podCreationTimestamp="2026-03-17 18:58:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:58:03.853723134 +0000 UTC m=+627.886334656" watchObservedRunningTime="2026-03-17 18:58:03.858411847 +0000 UTC m=+627.891023359" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867643 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-systemd-units\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867690 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-ovn\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867713 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-cni-netd\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867726 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-env-overrides\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867742 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-node-log\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867756 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-run-netns\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.867781 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-slash\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871688 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-ovnkube-script-lib\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871779 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871818 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-var-lib-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871865 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41913ab6-74f8-46b4-be58-95d799bc5365-ovn-node-metrics-cert\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871890 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-cni-bin\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871928 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-run-ovn-kubernetes\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871948 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871974 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-kubelet\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.871990 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-ovnkube-config\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872028 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-etc-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872051 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-log-socket\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872111 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9fnc\" (UniqueName: \"kubernetes.io/projected/41913ab6-74f8-46b4-be58-95d799bc5365-kube-api-access-d9fnc\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872154 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-systemd\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872231 5110 reconciler_common.go:299] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872246 5110 reconciler_common.go:299] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872260 5110 reconciler_common.go:299] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872271 5110 reconciler_common.go:299] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872283 5110 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872296 5110 reconciler_common.go:299] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872308 5110 reconciler_common.go:299] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872321 5110 reconciler_common.go:299] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872336 5110 reconciler_common.go:299] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872347 5110 reconciler_common.go:299] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872358 5110 reconciler_common.go:299] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-slash\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872368 5110 reconciler_common.go:299] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872378 5110 reconciler_common.go:299] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872388 5110 reconciler_common.go:299] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872400 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8b652\" (UniqueName: \"kubernetes.io/projected/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-kube-api-access-8b652\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872412 5110 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872423 5110 reconciler_common.go:299] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-node-log\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872435 5110 reconciler_common.go:299] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-log-socket\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872446 5110 reconciler_common.go:299] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.872457 5110 reconciler_common.go:299] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.879740 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4jphj"] Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.890708 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4jphj"] Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.904220 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488"] Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.906510 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-jr488"] Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.934241 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.944780 5110 scope.go:117] "RemoveContainer" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.977357 5110 scope.go:117] "RemoveContainer" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:03 crc kubenswrapper[5110]: E0317 18:58:03.977604 5110 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf20fbc8f_f15b_4713_9b57_aa5dbcb0d9b8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe3aacec_3e82_49dc_833e_4b1d848ea2c6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe3aacec_3e82_49dc_833e_4b1d848ea2c6.slice/crio-98f18290f31e5aaa7954e356a85a85b8e221a7703ec1956532614cb7401153aa\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf20fbc8f_f15b_4713_9b57_aa5dbcb0d9b8.slice/crio-d9fdf49e963657295a75c522c5fb9bc94198871a002c11e3280ef9eb82012cdc\": RecentStats: unable to find data in memory cache]" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.977812 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.977899 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-var-lib-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.977953 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41913ab6-74f8-46b4-be58-95d799bc5365-ovn-node-metrics-cert\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.977977 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-cni-bin\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978029 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-run-ovn-kubernetes\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978072 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978101 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-kubelet\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978120 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-ovnkube-config\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978165 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-etc-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978192 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-log-socket\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978221 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d9fnc\" (UniqueName: \"kubernetes.io/projected/41913ab6-74f8-46b4-be58-95d799bc5365-kube-api-access-d9fnc\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978244 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-systemd\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978273 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-kubelet\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978294 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-systemd-units\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978342 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-systemd-units\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978351 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-ovn\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978384 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-cni-netd\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978411 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-env-overrides\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978415 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-run-ovn-kubernetes\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978414 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-log-socket\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978437 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-node-log\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978461 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-run-netns\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978464 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978496 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-cni-bin\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978517 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-slash\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978551 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-ovnkube-script-lib\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978621 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-etc-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978711 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-var-lib-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978746 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-run-netns\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978775 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-node-log\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978962 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-cni-netd\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978973 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-systemd\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.978390 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-ovn\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.979031 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-host-slash\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.977962 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41913ab6-74f8-46b4-be58-95d799bc5365-run-openvswitch\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.980117 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-ovnkube-config\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.983744 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-env-overrides\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.983977 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41913ab6-74f8-46b4-be58-95d799bc5365-ovnkube-script-lib\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.987942 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41913ab6-74f8-46b4-be58-95d799bc5365-ovn-node-metrics-cert\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:03 crc kubenswrapper[5110]: I0317 18:58:03.999008 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9fnc\" (UniqueName: \"kubernetes.io/projected/41913ab6-74f8-46b4-be58-95d799bc5365-kube-api-access-d9fnc\") pod \"ovnkube-node-h6d87\" (UID: \"41913ab6-74f8-46b4-be58-95d799bc5365\") " pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.030617 5110 scope.go:117] "RemoveContainer" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.037467 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.053260 5110 scope.go:117] "RemoveContainer" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: W0317 18:58:04.061270 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41913ab6_74f8_46b4_be58_95d799bc5365.slice/crio-ef9e1693380234f9a1eff7b716c855d508a748f1fef460684d4cd2a10475c34d WatchSource:0}: Error finding container ef9e1693380234f9a1eff7b716c855d508a748f1fef460684d4cd2a10475c34d: Status 404 returned error can't find the container with id ef9e1693380234f9a1eff7b716c855d508a748f1fef460684d4cd2a10475c34d Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.079461 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24rm2\" (UniqueName: \"kubernetes.io/projected/3e77e6a5-f972-41ad-9d00-d86d43c0176d-kube-api-access-24rm2\") pod \"3e77e6a5-f972-41ad-9d00-d86d43c0176d\" (UID: \"3e77e6a5-f972-41ad-9d00-d86d43c0176d\") " Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.084741 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e77e6a5-f972-41ad-9d00-d86d43c0176d-kube-api-access-24rm2" (OuterVolumeSpecName: "kube-api-access-24rm2") pod "3e77e6a5-f972-41ad-9d00-d86d43c0176d" (UID: "3e77e6a5-f972-41ad-9d00-d86d43c0176d"). InnerVolumeSpecName "kube-api-access-24rm2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.092218 5110 scope.go:117] "RemoveContainer" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.108492 5110 scope.go:117] "RemoveContainer" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.127277 5110 scope.go:117] "RemoveContainer" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.144436 5110 scope.go:117] "RemoveContainer" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.145031 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": container with ID starting with 0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a not found: ID does not exist" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.145098 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} err="failed to get container status \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": rpc error: code = NotFound desc = could not find container \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": container with ID starting with 0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.145122 5110 scope.go:117] "RemoveContainer" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.145501 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": container with ID starting with 888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef not found: ID does not exist" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.145550 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} err="failed to get container status \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": rpc error: code = NotFound desc = could not find container \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": container with ID starting with 888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.145581 5110 scope.go:117] "RemoveContainer" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.145953 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": container with ID starting with 68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11 not found: ID does not exist" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.145989 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} err="failed to get container status \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": rpc error: code = NotFound desc = could not find container \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": container with ID starting with 68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.146017 5110 scope.go:117] "RemoveContainer" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.146399 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": container with ID starting with 94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f not found: ID does not exist" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.146431 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} err="failed to get container status \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": rpc error: code = NotFound desc = could not find container \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": container with ID starting with 94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.146450 5110 scope.go:117] "RemoveContainer" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.146730 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": container with ID starting with 04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290 not found: ID does not exist" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.146760 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} err="failed to get container status \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": rpc error: code = NotFound desc = could not find container \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": container with ID starting with 04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.146778 5110 scope.go:117] "RemoveContainer" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.147042 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": container with ID starting with 3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f not found: ID does not exist" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.147085 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} err="failed to get container status \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": rpc error: code = NotFound desc = could not find container \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": container with ID starting with 3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.147101 5110 scope.go:117] "RemoveContainer" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.147333 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": container with ID starting with 9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187 not found: ID does not exist" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.147364 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} err="failed to get container status \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": rpc error: code = NotFound desc = could not find container \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": container with ID starting with 9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.147396 5110 scope.go:117] "RemoveContainer" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.147646 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": container with ID starting with 9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9 not found: ID does not exist" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.147680 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} err="failed to get container status \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": rpc error: code = NotFound desc = could not find container \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": container with ID starting with 9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.147701 5110 scope.go:117] "RemoveContainer" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" Mar 17 18:58:04 crc kubenswrapper[5110]: E0317 18:58:04.148086 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": container with ID starting with b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a not found: ID does not exist" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.148192 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} err="failed to get container status \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": rpc error: code = NotFound desc = could not find container \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": container with ID starting with b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.148279 5110 scope.go:117] "RemoveContainer" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.148616 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} err="failed to get container status \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": rpc error: code = NotFound desc = could not find container \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": container with ID starting with 0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.148636 5110 scope.go:117] "RemoveContainer" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.149033 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} err="failed to get container status \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": rpc error: code = NotFound desc = could not find container \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": container with ID starting with 888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.149159 5110 scope.go:117] "RemoveContainer" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.149663 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} err="failed to get container status \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": rpc error: code = NotFound desc = could not find container \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": container with ID starting with 68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.149685 5110 scope.go:117] "RemoveContainer" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.149949 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} err="failed to get container status \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": rpc error: code = NotFound desc = could not find container \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": container with ID starting with 94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.149967 5110 scope.go:117] "RemoveContainer" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.150229 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} err="failed to get container status \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": rpc error: code = NotFound desc = could not find container \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": container with ID starting with 04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.150307 5110 scope.go:117] "RemoveContainer" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.150608 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} err="failed to get container status \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": rpc error: code = NotFound desc = could not find container \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": container with ID starting with 3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.150638 5110 scope.go:117] "RemoveContainer" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.150975 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} err="failed to get container status \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": rpc error: code = NotFound desc = could not find container \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": container with ID starting with 9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151013 5110 scope.go:117] "RemoveContainer" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151349 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} err="failed to get container status \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": rpc error: code = NotFound desc = could not find container \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": container with ID starting with 9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151430 5110 scope.go:117] "RemoveContainer" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151688 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} err="failed to get container status \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": rpc error: code = NotFound desc = could not find container \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": container with ID starting with b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151725 5110 scope.go:117] "RemoveContainer" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151907 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} err="failed to get container status \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": rpc error: code = NotFound desc = could not find container \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": container with ID starting with 0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.151944 5110 scope.go:117] "RemoveContainer" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.152114 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} err="failed to get container status \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": rpc error: code = NotFound desc = could not find container \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": container with ID starting with 888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.152129 5110 scope.go:117] "RemoveContainer" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.152327 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} err="failed to get container status \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": rpc error: code = NotFound desc = could not find container \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": container with ID starting with 68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.152454 5110 scope.go:117] "RemoveContainer" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.152830 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} err="failed to get container status \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": rpc error: code = NotFound desc = could not find container \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": container with ID starting with 94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.152851 5110 scope.go:117] "RemoveContainer" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.153160 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} err="failed to get container status \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": rpc error: code = NotFound desc = could not find container \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": container with ID starting with 04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.153249 5110 scope.go:117] "RemoveContainer" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.153519 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} err="failed to get container status \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": rpc error: code = NotFound desc = could not find container \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": container with ID starting with 3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.153562 5110 scope.go:117] "RemoveContainer" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.153995 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} err="failed to get container status \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": rpc error: code = NotFound desc = could not find container \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": container with ID starting with 9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.154101 5110 scope.go:117] "RemoveContainer" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.154416 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} err="failed to get container status \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": rpc error: code = NotFound desc = could not find container \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": container with ID starting with 9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.154457 5110 scope.go:117] "RemoveContainer" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.154986 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} err="failed to get container status \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": rpc error: code = NotFound desc = could not find container \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": container with ID starting with b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155023 5110 scope.go:117] "RemoveContainer" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155291 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} err="failed to get container status \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": rpc error: code = NotFound desc = could not find container \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": container with ID starting with 0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155336 5110 scope.go:117] "RemoveContainer" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155560 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} err="failed to get container status \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": rpc error: code = NotFound desc = could not find container \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": container with ID starting with 888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155603 5110 scope.go:117] "RemoveContainer" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155816 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} err="failed to get container status \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": rpc error: code = NotFound desc = could not find container \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": container with ID starting with 68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.155866 5110 scope.go:117] "RemoveContainer" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156086 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} err="failed to get container status \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": rpc error: code = NotFound desc = could not find container \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": container with ID starting with 94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156109 5110 scope.go:117] "RemoveContainer" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156344 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} err="failed to get container status \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": rpc error: code = NotFound desc = could not find container \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": container with ID starting with 04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156384 5110 scope.go:117] "RemoveContainer" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156597 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} err="failed to get container status \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": rpc error: code = NotFound desc = could not find container \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": container with ID starting with 3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156638 5110 scope.go:117] "RemoveContainer" containerID="9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156852 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187"} err="failed to get container status \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": rpc error: code = NotFound desc = could not find container \"9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187\": container with ID starting with 9d105693f93e13b5a47d62b6c5786be63f34dbcd61d7718a4569485179967187 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.156916 5110 scope.go:117] "RemoveContainer" containerID="9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.157285 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9"} err="failed to get container status \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": rpc error: code = NotFound desc = could not find container \"9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9\": container with ID starting with 9ebc052969fddb7a65464e208fa1381bb573b3562c31e9f624a8fab1eaf90cf9 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.157333 5110 scope.go:117] "RemoveContainer" containerID="b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.157700 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a"} err="failed to get container status \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": rpc error: code = NotFound desc = could not find container \"b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a\": container with ID starting with b8c5f75874be629b2b1707fd1ac8ad4c0eeb8ebf2106dc1d7acb2fe99ab1698a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.157741 5110 scope.go:117] "RemoveContainer" containerID="0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158016 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a"} err="failed to get container status \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": rpc error: code = NotFound desc = could not find container \"0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a\": container with ID starting with 0bed454d8025acab6467cf48e4b1f82d7c7d45961b1140d8c66d2177f713a46a not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158034 5110 scope.go:117] "RemoveContainer" containerID="888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158275 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef"} err="failed to get container status \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": rpc error: code = NotFound desc = could not find container \"888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef\": container with ID starting with 888b0755feda08598f6904733a147288677078944f0762baf3a8e6905f7545ef not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158309 5110 scope.go:117] "RemoveContainer" containerID="68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158525 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11"} err="failed to get container status \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": rpc error: code = NotFound desc = could not find container \"68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11\": container with ID starting with 68ad5382eb892da35c5eb62bc01358d4d80fbf563f5030ab43e68003c2840f11 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158544 5110 scope.go:117] "RemoveContainer" containerID="94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158809 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f"} err="failed to get container status \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": rpc error: code = NotFound desc = could not find container \"94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f\": container with ID starting with 94f4e8c350e1cad1812023e535036e6c426f7d2f9640cfd43093bea373b4d87f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.158827 5110 scope.go:117] "RemoveContainer" containerID="04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.159093 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290"} err="failed to get container status \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": rpc error: code = NotFound desc = could not find container \"04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290\": container with ID starting with 04b9ec28e1b7b05a2069ae5f34f99d68d59d9647976dee1bd75cd81ea7c92290 not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.159109 5110 scope.go:117] "RemoveContainer" containerID="3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.159369 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f"} err="failed to get container status \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": rpc error: code = NotFound desc = could not find container \"3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f\": container with ID starting with 3abf74dcda81812b0b9cd3d9164defaea0268579b9778eb9a462215603eeed9f not found: ID does not exist" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.181080 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-24rm2\" (UniqueName: \"kubernetes.io/projected/3e77e6a5-f972-41ad-9d00-d86d43c0176d-kube-api-access-24rm2\") on node \"crc\" DevicePath \"\"" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.762730 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8" path="/var/lib/kubelet/pods/f20fbc8f-f15b-4713-9b57-aa5dbcb0d9b8/volumes" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.773297 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe3aacec-3e82-49dc-833e-4b1d848ea2c6" path="/var/lib/kubelet/pods/fe3aacec-3e82-49dc-833e-4b1d848ea2c6/volumes" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.840447 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.840625 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6kh2q" event={"ID":"7346b312-05b5-4475-8d54-82709b69e7ed","Type":"ContainerStarted","Data":"503b1d64af83e519cae385a937e9eaa06a8fd034863608bec0f2c6cd7dce4ef6"} Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.845354 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" event={"ID":"3e77e6a5-f972-41ad-9d00-d86d43c0176d","Type":"ContainerDied","Data":"3d05d919ecba0416f9bf60ef9fa1a77eb705bc0bc1898685928d4152ae8c788f"} Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.845477 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d05d919ecba0416f9bf60ef9fa1a77eb705bc0bc1898685928d4152ae8c788f" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.845365 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562898-5lc6h" Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.847891 5110 generic.go:358] "Generic (PLEG): container finished" podID="41913ab6-74f8-46b4-be58-95d799bc5365" containerID="681d6ae0b616aee10047bc9a59e056d2e6b2ba84c848edb8b7c71d768205a44d" exitCode=0 Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.848033 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerDied","Data":"681d6ae0b616aee10047bc9a59e056d2e6b2ba84c848edb8b7c71d768205a44d"} Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.848111 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"ef9e1693380234f9a1eff7b716c855d508a748f1fef460684d4cd2a10475c34d"} Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.992238 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562892-2hdnb"] Mar 17 18:58:04 crc kubenswrapper[5110]: I0317 18:58:04.997235 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562892-2hdnb"] Mar 17 18:58:05 crc kubenswrapper[5110]: I0317 18:58:05.865255 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"64a3b1884d26badc190ebf7e0b82af69639662bd0568c017d4f3c8867ab597c4"} Mar 17 18:58:05 crc kubenswrapper[5110]: I0317 18:58:05.865316 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"8c2cca3201b1bb444de56aab0cf9b85901be9edd193af70894e1e133c8bf98c2"} Mar 17 18:58:05 crc kubenswrapper[5110]: I0317 18:58:05.865330 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"8c814c93e17937c4cde73e7a944f7833cfc489ea375b97d35050e2845e83b7b6"} Mar 17 18:58:05 crc kubenswrapper[5110]: I0317 18:58:05.865339 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"177b5979e66ca82b393ddbd5b1c19d1d1de664892b50a5d56efabe1539e96a2c"} Mar 17 18:58:05 crc kubenswrapper[5110]: I0317 18:58:05.865350 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"1269b9cdcbaa40b03bc2a624c0a6854009c18cbac36da248dc998c1e91730d8e"} Mar 17 18:58:05 crc kubenswrapper[5110]: I0317 18:58:05.865362 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"58168b598025aa6cf07563d8cf8a34645e32e9689316a601295e32013dfeb261"} Mar 17 18:58:06 crc kubenswrapper[5110]: I0317 18:58:06.752751 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64571774-f18c-4140-b48c-f8868af19eb7" path="/var/lib/kubelet/pods/64571774-f18c-4140-b48c-f8868af19eb7/volumes" Mar 17 18:58:08 crc kubenswrapper[5110]: I0317 18:58:08.887105 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"483cb70f0c36bbd304f69369dd90612aae0769b341a243bc6da5f9513122a2b0"} Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.909413 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" event={"ID":"41913ab6-74f8-46b4-be58-95d799bc5365","Type":"ContainerStarted","Data":"d6b3c047b706854e74c0ece078f485f4fa45661dbc5eab0e44d54655bcc4b988"} Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.910101 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.910290 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.910443 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.947516 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" podStartSLOduration=7.947491155 podStartE2EDuration="7.947491155s" podCreationTimestamp="2026-03-17 18:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:58:10.943806178 +0000 UTC m=+634.976417710" watchObservedRunningTime="2026-03-17 18:58:10.947491155 +0000 UTC m=+634.980102717" Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.953691 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:10 crc kubenswrapper[5110]: I0317 18:58:10.961195 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.132777 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.133291 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.133358 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.134190 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c003f768dfcdd99f40db6c78311c785574f0d09dbaf1d45833991c2b96d1903"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.134289 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://6c003f768dfcdd99f40db6c78311c785574f0d09dbaf1d45833991c2b96d1903" gracePeriod=600 Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.259687 5110 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.924409 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="6c003f768dfcdd99f40db6c78311c785574f0d09dbaf1d45833991c2b96d1903" exitCode=0 Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.924487 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"6c003f768dfcdd99f40db6c78311c785574f0d09dbaf1d45833991c2b96d1903"} Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.925193 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"399b37230100df387aa59a5f7caddd2a37e880d366524aadcd930dd9995f833d"} Mar 17 18:58:12 crc kubenswrapper[5110]: I0317 18:58:12.925215 5110 scope.go:117] "RemoveContainer" containerID="ebe54d3668f3ba6d72d2f6801190bcc6cdfdce7f9ffcdbc79df03266e49b4ac1" Mar 17 18:58:37 crc kubenswrapper[5110]: I0317 18:58:37.333922 5110 scope.go:117] "RemoveContainer" containerID="dae6ed28016c0b928ee8326b119f77e0162442c304020b1acb272119720775e7" Mar 17 18:58:37 crc kubenswrapper[5110]: I0317 18:58:37.358824 5110 scope.go:117] "RemoveContainer" containerID="1d9b6064bf5bf722c17e500956c3fc8e9e1d29cc3c5d5d857368e4735bd28862" Mar 17 18:58:37 crc kubenswrapper[5110]: I0317 18:58:37.376741 5110 scope.go:117] "RemoveContainer" containerID="fbb91fd8cdbb67bf5e4ebf66b4eb5d69dfc2a8a7933accda111fb00479faa00b" Mar 17 18:58:42 crc kubenswrapper[5110]: I0317 18:58:42.956188 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h6d87" Mar 17 18:59:03 crc kubenswrapper[5110]: I0317 18:59:03.801328 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-krsp5"] Mar 17 18:59:03 crc kubenswrapper[5110]: I0317 18:59:03.802391 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-krsp5" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="registry-server" containerID="cri-o://91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" gracePeriod=30 Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.015090 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67 is running failed: container process not found" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.015956 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67 is running failed: container process not found" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.016495 5110 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67 is running failed: container process not found" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.016544 5110 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-krsp5" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="registry-server" probeResult="unknown" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.126191 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.197723 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-utilities\") pod \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.197773 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-catalog-content\") pod \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.197835 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-kube-api-access-fkrrs\") pod \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\" (UID: \"fc56a95b-a15c-4661-8247-0ec3d89ac2b6\") " Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.199292 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-utilities" (OuterVolumeSpecName: "utilities") pod "fc56a95b-a15c-4661-8247-0ec3d89ac2b6" (UID: "fc56a95b-a15c-4661-8247-0ec3d89ac2b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.203822 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-kube-api-access-fkrrs" (OuterVolumeSpecName: "kube-api-access-fkrrs") pod "fc56a95b-a15c-4661-8247-0ec3d89ac2b6" (UID: "fc56a95b-a15c-4661-8247-0ec3d89ac2b6"). InnerVolumeSpecName "kube-api-access-fkrrs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.215843 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc56a95b-a15c-4661-8247-0ec3d89ac2b6" (UID: "fc56a95b-a15c-4661-8247-0ec3d89ac2b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.282146 5110 generic.go:358] "Generic (PLEG): container finished" podID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" exitCode=0 Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.282213 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krsp5" event={"ID":"fc56a95b-a15c-4661-8247-0ec3d89ac2b6","Type":"ContainerDied","Data":"91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67"} Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.282264 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krsp5" event={"ID":"fc56a95b-a15c-4661-8247-0ec3d89ac2b6","Type":"ContainerDied","Data":"d9c88f9e465b8429f90d6a0de781e7a506d2786144da9d47a162df6017f81476"} Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.282283 5110 scope.go:117] "RemoveContainer" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.282343 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krsp5" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.297098 5110 scope.go:117] "RemoveContainer" containerID="f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.299244 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-kube-api-access-fkrrs\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.299270 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.299279 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc56a95b-a15c-4661-8247-0ec3d89ac2b6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.319114 5110 scope.go:117] "RemoveContainer" containerID="c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.337595 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-krsp5"] Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.342718 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-krsp5"] Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.344169 5110 scope.go:117] "RemoveContainer" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.344580 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67\": container with ID starting with 91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67 not found: ID does not exist" containerID="91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.344624 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67"} err="failed to get container status \"91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67\": rpc error: code = NotFound desc = could not find container \"91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67\": container with ID starting with 91823e26fd1f971ea7ef481379a51b4f766c6c893e4021106ffd7d10411e6b67 not found: ID does not exist" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.344647 5110 scope.go:117] "RemoveContainer" containerID="f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402" Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.345598 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402\": container with ID starting with f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402 not found: ID does not exist" containerID="f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.345633 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402"} err="failed to get container status \"f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402\": rpc error: code = NotFound desc = could not find container \"f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402\": container with ID starting with f3a2862132b7f5146a76c42b7584ff7c83cc162bbb56f1321d216eca6157f402 not found: ID does not exist" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.345654 5110 scope.go:117] "RemoveContainer" containerID="c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea" Mar 17 18:59:04 crc kubenswrapper[5110]: E0317 18:59:04.345948 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea\": container with ID starting with c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea not found: ID does not exist" containerID="c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.345985 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea"} err="failed to get container status \"c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea\": rpc error: code = NotFound desc = could not find container \"c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea\": container with ID starting with c1bc14fa457fb699c23080aa132e487f3f0bb1f7a5678aa13dfec212f49b53ea not found: ID does not exist" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.734325 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-d2vp6"] Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735011 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="extract-content" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735034 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="extract-content" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735076 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="extract-utilities" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735085 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="extract-utilities" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735095 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3e77e6a5-f972-41ad-9d00-d86d43c0176d" containerName="oc" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735102 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e77e6a5-f972-41ad-9d00-d86d43c0176d" containerName="oc" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735136 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="registry-server" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735144 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="registry-server" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735249 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="3e77e6a5-f972-41ad-9d00-d86d43c0176d" containerName="oc" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.735262 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" containerName="registry-server" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.748123 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.756501 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc56a95b-a15c-4661-8247-0ec3d89ac2b6" path="/var/lib/kubelet/pods/fc56a95b-a15c-4661-8247-0ec3d89ac2b6/volumes" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.757000 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-d2vp6"] Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.805635 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e81eed8-2208-4fcb-8829-43bd93af566a-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.805703 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e81eed8-2208-4fcb-8829-43bd93af566a-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.805786 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6drm\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-kube-api-access-h6drm\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.805884 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-bound-sa-token\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.805964 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e81eed8-2208-4fcb-8829-43bd93af566a-registry-certificates\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.806067 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e81eed8-2208-4fcb-8829-43bd93af566a-trusted-ca\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.806140 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.806208 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-registry-tls\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.830529 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907073 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-bound-sa-token\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907134 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e81eed8-2208-4fcb-8829-43bd93af566a-registry-certificates\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907173 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e81eed8-2208-4fcb-8829-43bd93af566a-trusted-ca\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907205 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-registry-tls\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907240 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e81eed8-2208-4fcb-8829-43bd93af566a-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907267 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e81eed8-2208-4fcb-8829-43bd93af566a-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.907288 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-h6drm\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-kube-api-access-h6drm\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.908368 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e81eed8-2208-4fcb-8829-43bd93af566a-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.908716 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e81eed8-2208-4fcb-8829-43bd93af566a-trusted-ca\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.909607 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e81eed8-2208-4fcb-8829-43bd93af566a-registry-certificates\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.914731 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e81eed8-2208-4fcb-8829-43bd93af566a-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.915575 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-registry-tls\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.925645 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-bound-sa-token\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:04 crc kubenswrapper[5110]: I0317 18:59:04.927105 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6drm\" (UniqueName: \"kubernetes.io/projected/0e81eed8-2208-4fcb-8829-43bd93af566a-kube-api-access-h6drm\") pod \"image-registry-5d9d95bf5b-d2vp6\" (UID: \"0e81eed8-2208-4fcb-8829-43bd93af566a\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:05 crc kubenswrapper[5110]: I0317 18:59:05.065897 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:05 crc kubenswrapper[5110]: I0317 18:59:05.286535 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-d2vp6"] Mar 17 18:59:06 crc kubenswrapper[5110]: I0317 18:59:06.300981 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" event={"ID":"0e81eed8-2208-4fcb-8829-43bd93af566a","Type":"ContainerStarted","Data":"9645ec3f9f42c05bfa21e2193e9785bdf6a4e4aad98f76fb003f60f8c13104a9"} Mar 17 18:59:06 crc kubenswrapper[5110]: I0317 18:59:06.301323 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:06 crc kubenswrapper[5110]: I0317 18:59:06.301334 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" event={"ID":"0e81eed8-2208-4fcb-8829-43bd93af566a","Type":"ContainerStarted","Data":"d66ee425a79ce36c90380c3aca1f9a934e4c8d0e9fbf09867907511dd6cf9395"} Mar 17 18:59:06 crc kubenswrapper[5110]: I0317 18:59:06.322428 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" podStartSLOduration=2.3224117890000002 podStartE2EDuration="2.322411789s" podCreationTimestamp="2026-03-17 18:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 18:59:06.319639698 +0000 UTC m=+690.352251220" watchObservedRunningTime="2026-03-17 18:59:06.322411789 +0000 UTC m=+690.355023311" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.428835 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb"] Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.451930 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb"] Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.452094 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.455421 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.542723 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.542916 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58722\" (UniqueName: \"kubernetes.io/projected/738d9d50-28e2-4982-bee6-7382cff35aaa-kube-api-access-58722\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.542949 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.644466 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.644574 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-58722\" (UniqueName: \"kubernetes.io/projected/738d9d50-28e2-4982-bee6-7382cff35aaa-kube-api-access-58722\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.644601 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.645222 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.645316 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.669269 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-58722\" (UniqueName: \"kubernetes.io/projected/738d9d50-28e2-4982-bee6-7382cff35aaa-kube-api-access-58722\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.771780 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:07 crc kubenswrapper[5110]: I0317 18:59:07.942280 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb"] Mar 17 18:59:08 crc kubenswrapper[5110]: I0317 18:59:08.315943 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" event={"ID":"738d9d50-28e2-4982-bee6-7382cff35aaa","Type":"ContainerStarted","Data":"d478a743608b967a7474a0ee72f36a06431c1f9f4981073e64088c2a14a105a5"} Mar 17 18:59:08 crc kubenswrapper[5110]: I0317 18:59:08.315996 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" event={"ID":"738d9d50-28e2-4982-bee6-7382cff35aaa","Type":"ContainerStarted","Data":"78485b81f3b7eac8b983b80218f2ebb5846161e51f64d09fb392c039b77d6cb6"} Mar 17 18:59:09 crc kubenswrapper[5110]: I0317 18:59:09.321327 5110 generic.go:358] "Generic (PLEG): container finished" podID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerID="d478a743608b967a7474a0ee72f36a06431c1f9f4981073e64088c2a14a105a5" exitCode=0 Mar 17 18:59:09 crc kubenswrapper[5110]: I0317 18:59:09.321471 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" event={"ID":"738d9d50-28e2-4982-bee6-7382cff35aaa","Type":"ContainerDied","Data":"d478a743608b967a7474a0ee72f36a06431c1f9f4981073e64088c2a14a105a5"} Mar 17 18:59:11 crc kubenswrapper[5110]: I0317 18:59:11.337151 5110 generic.go:358] "Generic (PLEG): container finished" podID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerID="abc6c095670ac39765ef8cc1648df887ab034190ce01a99d88e2637d0b774f62" exitCode=0 Mar 17 18:59:11 crc kubenswrapper[5110]: I0317 18:59:11.337269 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" event={"ID":"738d9d50-28e2-4982-bee6-7382cff35aaa","Type":"ContainerDied","Data":"abc6c095670ac39765ef8cc1648df887ab034190ce01a99d88e2637d0b774f62"} Mar 17 18:59:12 crc kubenswrapper[5110]: I0317 18:59:12.348906 5110 generic.go:358] "Generic (PLEG): container finished" podID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerID="29e614302840a326fafbdc07ab41c3e2fd1c09b196f009c8244b7a999a323ec8" exitCode=0 Mar 17 18:59:12 crc kubenswrapper[5110]: I0317 18:59:12.348995 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" event={"ID":"738d9d50-28e2-4982-bee6-7382cff35aaa","Type":"ContainerDied","Data":"29e614302840a326fafbdc07ab41c3e2fd1c09b196f009c8244b7a999a323ec8"} Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.594111 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.726553 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-bundle\") pod \"738d9d50-28e2-4982-bee6-7382cff35aaa\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.726895 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58722\" (UniqueName: \"kubernetes.io/projected/738d9d50-28e2-4982-bee6-7382cff35aaa-kube-api-access-58722\") pod \"738d9d50-28e2-4982-bee6-7382cff35aaa\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.726933 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-util\") pod \"738d9d50-28e2-4982-bee6-7382cff35aaa\" (UID: \"738d9d50-28e2-4982-bee6-7382cff35aaa\") " Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.733240 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-bundle" (OuterVolumeSpecName: "bundle") pod "738d9d50-28e2-4982-bee6-7382cff35aaa" (UID: "738d9d50-28e2-4982-bee6-7382cff35aaa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.735090 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/738d9d50-28e2-4982-bee6-7382cff35aaa-kube-api-access-58722" (OuterVolumeSpecName: "kube-api-access-58722") pod "738d9d50-28e2-4982-bee6-7382cff35aaa" (UID: "738d9d50-28e2-4982-bee6-7382cff35aaa"). InnerVolumeSpecName "kube-api-access-58722". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.736975 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-util" (OuterVolumeSpecName: "util") pod "738d9d50-28e2-4982-bee6-7382cff35aaa" (UID: "738d9d50-28e2-4982-bee6-7382cff35aaa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.828017 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-58722\" (UniqueName: \"kubernetes.io/projected/738d9d50-28e2-4982-bee6-7382cff35aaa-kube-api-access-58722\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.828076 5110 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-util\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:13 crc kubenswrapper[5110]: I0317 18:59:13.828088 5110 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/738d9d50-28e2-4982-bee6-7382cff35aaa-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.365649 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" event={"ID":"738d9d50-28e2-4982-bee6-7382cff35aaa","Type":"ContainerDied","Data":"78485b81f3b7eac8b983b80218f2ebb5846161e51f64d09fb392c039b77d6cb6"} Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.365708 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78485b81f3b7eac8b983b80218f2ebb5846161e51f64d09fb392c039b77d6cb6" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.365670 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.625728 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f"] Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626336 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="pull" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626352 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="pull" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626376 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="extract" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626382 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="extract" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626391 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="util" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626397 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="util" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.626477 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="738d9d50-28e2-4982-bee6-7382cff35aaa" containerName="extract" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.636326 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.641161 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.644136 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f"] Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.742162 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.742361 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fjvm\" (UniqueName: \"kubernetes.io/projected/74637df8-ebd0-4afa-b472-12067b6b46c7-kube-api-access-7fjvm\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.742430 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.843637 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.843888 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.843977 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7fjvm\" (UniqueName: \"kubernetes.io/projected/74637df8-ebd0-4afa-b472-12067b6b46c7-kube-api-access-7fjvm\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.844311 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.844507 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.869176 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fjvm\" (UniqueName: \"kubernetes.io/projected/74637df8-ebd0-4afa-b472-12067b6b46c7-kube-api-access-7fjvm\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:14 crc kubenswrapper[5110]: I0317 18:59:14.963621 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.271321 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f"] Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.374566 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" event={"ID":"74637df8-ebd0-4afa-b472-12067b6b46c7","Type":"ContainerStarted","Data":"8a81a043f65099a441e974a2053fd57f6e883de6936cb67bc9129ed52918a703"} Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.415447 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8"] Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.422853 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.427176 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8"] Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.553949 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.554017 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.554045 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mhbg\" (UniqueName: \"kubernetes.io/projected/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-kube-api-access-4mhbg\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.655710 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.655763 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.655796 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4mhbg\" (UniqueName: \"kubernetes.io/projected/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-kube-api-access-4mhbg\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.656352 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.656433 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.673602 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mhbg\" (UniqueName: \"kubernetes.io/projected/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-kube-api-access-4mhbg\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:15 crc kubenswrapper[5110]: I0317 18:59:15.739728 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:16 crc kubenswrapper[5110]: I0317 18:59:16.133914 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8"] Mar 17 18:59:16 crc kubenswrapper[5110]: I0317 18:59:16.381653 5110 generic.go:358] "Generic (PLEG): container finished" podID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerID="0d5e2dbbdac0801d91b9efa54b79e01cb5d6d3b818159a9089652a7f645a744e" exitCode=0 Mar 17 18:59:16 crc kubenswrapper[5110]: I0317 18:59:16.382022 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" event={"ID":"74637df8-ebd0-4afa-b472-12067b6b46c7","Type":"ContainerDied","Data":"0d5e2dbbdac0801d91b9efa54b79e01cb5d6d3b818159a9089652a7f645a744e"} Mar 17 18:59:16 crc kubenswrapper[5110]: I0317 18:59:16.384412 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" event={"ID":"8e42598c-1b8a-48bb-a6cb-58a5e0f703da","Type":"ContainerStarted","Data":"fa9b2d96a394c5dd3be61209364bec1bc048c5375d200380203558f6a3a0a06e"} Mar 17 18:59:16 crc kubenswrapper[5110]: I0317 18:59:16.384462 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" event={"ID":"8e42598c-1b8a-48bb-a6cb-58a5e0f703da","Type":"ContainerStarted","Data":"2c4abf3a76e46bd0514c7c379985ccf2bd8f599b77746beaba5ddfcc2a6826a2"} Mar 17 18:59:17 crc kubenswrapper[5110]: I0317 18:59:17.396081 5110 generic.go:358] "Generic (PLEG): container finished" podID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerID="fa9b2d96a394c5dd3be61209364bec1bc048c5375d200380203558f6a3a0a06e" exitCode=0 Mar 17 18:59:17 crc kubenswrapper[5110]: I0317 18:59:17.396253 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" event={"ID":"8e42598c-1b8a-48bb-a6cb-58a5e0f703da","Type":"ContainerDied","Data":"fa9b2d96a394c5dd3be61209364bec1bc048c5375d200380203558f6a3a0a06e"} Mar 17 18:59:18 crc kubenswrapper[5110]: I0317 18:59:18.403476 5110 generic.go:358] "Generic (PLEG): container finished" podID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerID="a00d14619be0f8c95a75a3dff8e4b599297741764a8b3bd6d2bdb33dd1efadca" exitCode=0 Mar 17 18:59:18 crc kubenswrapper[5110]: I0317 18:59:18.403757 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" event={"ID":"74637df8-ebd0-4afa-b472-12067b6b46c7","Type":"ContainerDied","Data":"a00d14619be0f8c95a75a3dff8e4b599297741764a8b3bd6d2bdb33dd1efadca"} Mar 17 18:59:18 crc kubenswrapper[5110]: I0317 18:59:18.406910 5110 generic.go:358] "Generic (PLEG): container finished" podID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerID="3a2c9ba6b52fcb775274c89bc384707730d9591665fd2eba5733840ec7ad8827" exitCode=0 Mar 17 18:59:18 crc kubenswrapper[5110]: I0317 18:59:18.407028 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" event={"ID":"8e42598c-1b8a-48bb-a6cb-58a5e0f703da","Type":"ContainerDied","Data":"3a2c9ba6b52fcb775274c89bc384707730d9591665fd2eba5733840ec7ad8827"} Mar 17 18:59:19 crc kubenswrapper[5110]: I0317 18:59:19.415411 5110 generic.go:358] "Generic (PLEG): container finished" podID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerID="d4895d6f113c6e20b7dcc9c8c83e9e4dec479de447c5cc6cc182e767e9a16a6e" exitCode=0 Mar 17 18:59:19 crc kubenswrapper[5110]: I0317 18:59:19.415455 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" event={"ID":"74637df8-ebd0-4afa-b472-12067b6b46c7","Type":"ContainerDied","Data":"d4895d6f113c6e20b7dcc9c8c83e9e4dec479de447c5cc6cc182e767e9a16a6e"} Mar 17 18:59:19 crc kubenswrapper[5110]: I0317 18:59:19.418338 5110 generic.go:358] "Generic (PLEG): container finished" podID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerID="7a82d41846afc53f7cded706325903267c9f43aaf5f7cfc09c95ed963d72881e" exitCode=0 Mar 17 18:59:19 crc kubenswrapper[5110]: I0317 18:59:19.418383 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" event={"ID":"8e42598c-1b8a-48bb-a6cb-58a5e0f703da","Type":"ContainerDied","Data":"7a82d41846afc53f7cded706325903267c9f43aaf5f7cfc09c95ed963d72881e"} Mar 17 18:59:20 crc kubenswrapper[5110]: I0317 18:59:20.844330 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:20 crc kubenswrapper[5110]: I0317 18:59:20.876261 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.033700 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fjvm\" (UniqueName: \"kubernetes.io/projected/74637df8-ebd0-4afa-b472-12067b6b46c7-kube-api-access-7fjvm\") pod \"74637df8-ebd0-4afa-b472-12067b6b46c7\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.033781 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-util\") pod \"74637df8-ebd0-4afa-b472-12067b6b46c7\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.033821 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-util\") pod \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.033859 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-bundle\") pod \"74637df8-ebd0-4afa-b472-12067b6b46c7\" (UID: \"74637df8-ebd0-4afa-b472-12067b6b46c7\") " Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.033891 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mhbg\" (UniqueName: \"kubernetes.io/projected/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-kube-api-access-4mhbg\") pod \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.033947 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-bundle\") pod \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\" (UID: \"8e42598c-1b8a-48bb-a6cb-58a5e0f703da\") " Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.035011 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-bundle" (OuterVolumeSpecName: "bundle") pod "8e42598c-1b8a-48bb-a6cb-58a5e0f703da" (UID: "8e42598c-1b8a-48bb-a6cb-58a5e0f703da"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.036154 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-bundle" (OuterVolumeSpecName: "bundle") pod "74637df8-ebd0-4afa-b472-12067b6b46c7" (UID: "74637df8-ebd0-4afa-b472-12067b6b46c7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.044573 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74637df8-ebd0-4afa-b472-12067b6b46c7-kube-api-access-7fjvm" (OuterVolumeSpecName: "kube-api-access-7fjvm") pod "74637df8-ebd0-4afa-b472-12067b6b46c7" (UID: "74637df8-ebd0-4afa-b472-12067b6b46c7"). InnerVolumeSpecName "kube-api-access-7fjvm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.044612 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-kube-api-access-4mhbg" (OuterVolumeSpecName: "kube-api-access-4mhbg") pod "8e42598c-1b8a-48bb-a6cb-58a5e0f703da" (UID: "8e42598c-1b8a-48bb-a6cb-58a5e0f703da"). InnerVolumeSpecName "kube-api-access-4mhbg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.056276 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-util" (OuterVolumeSpecName: "util") pod "8e42598c-1b8a-48bb-a6cb-58a5e0f703da" (UID: "8e42598c-1b8a-48bb-a6cb-58a5e0f703da"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.057514 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-util" (OuterVolumeSpecName: "util") pod "74637df8-ebd0-4afa-b472-12067b6b46c7" (UID: "74637df8-ebd0-4afa-b472-12067b6b46c7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.135704 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7fjvm\" (UniqueName: \"kubernetes.io/projected/74637df8-ebd0-4afa-b472-12067b6b46c7-kube-api-access-7fjvm\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.135757 5110 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-util\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.135768 5110 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-util\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.135781 5110 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74637df8-ebd0-4afa-b472-12067b6b46c7-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.135793 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4mhbg\" (UniqueName: \"kubernetes.io/projected/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-kube-api-access-4mhbg\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.135804 5110 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e42598c-1b8a-48bb-a6cb-58a5e0f703da-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.431684 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" event={"ID":"74637df8-ebd0-4afa-b472-12067b6b46c7","Type":"ContainerDied","Data":"8a81a043f65099a441e974a2053fd57f6e883de6936cb67bc9129ed52918a703"} Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.431731 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a81a043f65099a441e974a2053fd57f6e883de6936cb67bc9129ed52918a703" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.431912 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.440692 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" event={"ID":"8e42598c-1b8a-48bb-a6cb-58a5e0f703da","Type":"ContainerDied","Data":"2c4abf3a76e46bd0514c7c379985ccf2bd8f599b77746beaba5ddfcc2a6826a2"} Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.440902 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c4abf3a76e46bd0514c7c379985ccf2bd8f599b77746beaba5ddfcc2a6826a2" Mar 17 18:59:21 crc kubenswrapper[5110]: I0317 18:59:21.440725 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.624088 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br"] Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625109 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="pull" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625128 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="pull" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625149 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="pull" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625157 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="pull" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625164 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="extract" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625172 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="extract" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625184 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="util" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625191 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="util" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625215 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="util" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625223 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="util" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625238 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="extract" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625245 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="extract" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625360 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="8e42598c-1b8a-48bb-a6cb-58a5e0f703da" containerName="extract" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.625375 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="74637df8-ebd0-4afa-b472-12067b6b46c7" containerName="extract" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.635785 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.637270 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br"] Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.638187 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.768173 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7lwh\" (UniqueName: \"kubernetes.io/projected/28ec609b-6d68-482d-8ce4-2a21d490f76a-kube-api-access-b7lwh\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.768256 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.768285 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.869414 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b7lwh\" (UniqueName: \"kubernetes.io/projected/28ec609b-6d68-482d-8ce4-2a21d490f76a-kube-api-access-b7lwh\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.869504 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.869538 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.870113 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.870336 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.891606 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7lwh\" (UniqueName: \"kubernetes.io/projected/28ec609b-6d68-482d-8ce4-2a21d490f76a-kube-api-access-b7lwh\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:23 crc kubenswrapper[5110]: I0317 18:59:23.959035 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:24 crc kubenswrapper[5110]: I0317 18:59:24.382963 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br"] Mar 17 18:59:24 crc kubenswrapper[5110]: W0317 18:59:24.393500 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28ec609b_6d68_482d_8ce4_2a21d490f76a.slice/crio-1552ab4b362c66e9f4431b36012e36c2241d7be22c1dc5a8343ad3839a54dd22 WatchSource:0}: Error finding container 1552ab4b362c66e9f4431b36012e36c2241d7be22c1dc5a8343ad3839a54dd22: Status 404 returned error can't find the container with id 1552ab4b362c66e9f4431b36012e36c2241d7be22c1dc5a8343ad3839a54dd22 Mar 17 18:59:24 crc kubenswrapper[5110]: I0317 18:59:24.482435 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" event={"ID":"28ec609b-6d68-482d-8ce4-2a21d490f76a","Type":"ContainerStarted","Data":"1552ab4b362c66e9f4431b36012e36c2241d7be22c1dc5a8343ad3839a54dd22"} Mar 17 18:59:25 crc kubenswrapper[5110]: I0317 18:59:25.488849 5110 generic.go:358] "Generic (PLEG): container finished" podID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerID="68874ed11a0a2728f1e84622a7fe0b19c400109870f84d14cab630f795864e29" exitCode=0 Mar 17 18:59:25 crc kubenswrapper[5110]: I0317 18:59:25.488911 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" event={"ID":"28ec609b-6d68-482d-8ce4-2a21d490f76a","Type":"ContainerDied","Data":"68874ed11a0a2728f1e84622a7fe0b19c400109870f84d14cab630f795864e29"} Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.312121 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-5d9d95bf5b-d2vp6" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.380988 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-5hkdz"] Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.582934 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm"] Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.615976 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm"] Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.616208 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.623267 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operators\"/\"kube-root-ca.crt\"" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.648435 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operators\"/\"openshift-service-ca.crt\"" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.648461 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-dockercfg-xzqtl\"" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.726230 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzwxv\" (UniqueName: \"kubernetes.io/projected/8cb83be5-bf66-476c-b29e-f055352697e2-kube-api-access-hzwxv\") pod \"obo-prometheus-operator-55568fc96c-6pbtm\" (UID: \"8cb83be5-bf66-476c-b29e-f055352697e2\") " pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.827835 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hzwxv\" (UniqueName: \"kubernetes.io/projected/8cb83be5-bf66-476c-b29e-f055352697e2-kube-api-access-hzwxv\") pod \"obo-prometheus-operator-55568fc96c-6pbtm\" (UID: \"8cb83be5-bf66-476c-b29e-f055352697e2\") " pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.880609 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzwxv\" (UniqueName: \"kubernetes.io/projected/8cb83be5-bf66-476c-b29e-f055352697e2-kube-api-access-hzwxv\") pod \"obo-prometheus-operator-55568fc96c-6pbtm\" (UID: \"8cb83be5-bf66-476c-b29e-f055352697e2\") " pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.949766 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.951256 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l"] Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.968299 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx"] Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.968543 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.972160 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-dockercfg-slh6n\"" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.972991 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l"] Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.973127 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.973429 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-service-cert\"" Mar 17 18:59:27 crc kubenswrapper[5110]: I0317 18:59:27.990198 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx"] Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.030023 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccfae76d-51c9-4f8c-a694-72d770811128-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-96h7l\" (UID: \"ccfae76d-51c9-4f8c-a694-72d770811128\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.030152 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/391b404f-2c22-46cf-954e-84540c6fe899-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx\" (UID: \"391b404f-2c22-46cf-954e-84540c6fe899\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.030205 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/391b404f-2c22-46cf-954e-84540c6fe899-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx\" (UID: \"391b404f-2c22-46cf-954e-84540c6fe899\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.030419 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccfae76d-51c9-4f8c-a694-72d770811128-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-96h7l\" (UID: \"ccfae76d-51c9-4f8c-a694-72d770811128\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.132190 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/391b404f-2c22-46cf-954e-84540c6fe899-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx\" (UID: \"391b404f-2c22-46cf-954e-84540c6fe899\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.132240 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/391b404f-2c22-46cf-954e-84540c6fe899-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx\" (UID: \"391b404f-2c22-46cf-954e-84540c6fe899\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.132338 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccfae76d-51c9-4f8c-a694-72d770811128-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-96h7l\" (UID: \"ccfae76d-51c9-4f8c-a694-72d770811128\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.132372 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccfae76d-51c9-4f8c-a694-72d770811128-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-96h7l\" (UID: \"ccfae76d-51c9-4f8c-a694-72d770811128\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.138891 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccfae76d-51c9-4f8c-a694-72d770811128-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-96h7l\" (UID: \"ccfae76d-51c9-4f8c-a694-72d770811128\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.139531 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/391b404f-2c22-46cf-954e-84540c6fe899-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx\" (UID: \"391b404f-2c22-46cf-954e-84540c6fe899\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.141199 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/391b404f-2c22-46cf-954e-84540c6fe899-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx\" (UID: \"391b404f-2c22-46cf-954e-84540c6fe899\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.141642 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccfae76d-51c9-4f8c-a694-72d770811128-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646cb684f4-96h7l\" (UID: \"ccfae76d-51c9-4f8c-a694-72d770811128\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.293037 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.306221 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.307148 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-587f9c8867-p6ncg"] Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.316709 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.320770 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"observability-operator-tls\"" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.321162 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"observability-operator-sa-dockercfg-nkjn8\"" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.344811 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-587f9c8867-p6ncg"] Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.436646 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn7b5\" (UniqueName: \"kubernetes.io/projected/0a812097-b357-4b5e-bfa5-115b26efc4e3-kube-api-access-kn7b5\") pod \"observability-operator-587f9c8867-p6ncg\" (UID: \"0a812097-b357-4b5e-bfa5-115b26efc4e3\") " pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.436713 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a812097-b357-4b5e-bfa5-115b26efc4e3-observability-operator-tls\") pod \"observability-operator-587f9c8867-p6ncg\" (UID: \"0a812097-b357-4b5e-bfa5-115b26efc4e3\") " pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.537753 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kn7b5\" (UniqueName: \"kubernetes.io/projected/0a812097-b357-4b5e-bfa5-115b26efc4e3-kube-api-access-kn7b5\") pod \"observability-operator-587f9c8867-p6ncg\" (UID: \"0a812097-b357-4b5e-bfa5-115b26efc4e3\") " pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.537811 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a812097-b357-4b5e-bfa5-115b26efc4e3-observability-operator-tls\") pod \"observability-operator-587f9c8867-p6ncg\" (UID: \"0a812097-b357-4b5e-bfa5-115b26efc4e3\") " pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.541368 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a812097-b357-4b5e-bfa5-115b26efc4e3-observability-operator-tls\") pod \"observability-operator-587f9c8867-p6ncg\" (UID: \"0a812097-b357-4b5e-bfa5-115b26efc4e3\") " pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.561403 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn7b5\" (UniqueName: \"kubernetes.io/projected/0a812097-b357-4b5e-bfa5-115b26efc4e3-kube-api-access-kn7b5\") pod \"observability-operator-587f9c8867-p6ncg\" (UID: \"0a812097-b357-4b5e-bfa5-115b26efc4e3\") " pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.633805 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.695634 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5b659947cf-krgf9"] Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.703259 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.706190 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"perses-operator-service-cert\"" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.707748 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5b659947cf-krgf9"] Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.709820 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"perses-operator-dockercfg-d62zc\"" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.741145 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdbdp\" (UniqueName: \"kubernetes.io/projected/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-kube-api-access-cdbdp\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.741193 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-webhook-cert\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.741280 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-apiservice-cert\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.741306 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-openshift-service-ca\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.842864 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-openshift-service-ca\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.842998 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cdbdp\" (UniqueName: \"kubernetes.io/projected/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-kube-api-access-cdbdp\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.843071 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-webhook-cert\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.843125 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-apiservice-cert\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.843643 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-openshift-service-ca\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.847534 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-apiservice-cert\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.847752 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-webhook-cert\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:28 crc kubenswrapper[5110]: I0317 18:59:28.862634 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdbdp\" (UniqueName: \"kubernetes.io/projected/c9354dcb-0033-473e-a3f3-e01e7b35d9eb-kube-api-access-cdbdp\") pod \"perses-operator-5b659947cf-krgf9\" (UID: \"c9354dcb-0033-473e-a3f3-e01e7b35d9eb\") " pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:29 crc kubenswrapper[5110]: I0317 18:59:29.021205 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.189530 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.313130 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.317756 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.345161 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-587f9c8867-p6ncg"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.446565 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5b659947cf-krgf9"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.515657 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5b659947cf-krgf9" event={"ID":"c9354dcb-0033-473e-a3f3-e01e7b35d9eb","Type":"ContainerStarted","Data":"594ffead259ab57b1c98bb5a537dc5b1b0e8ee73d5ae6a19109d64a0aa37aac3"} Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.516794 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" event={"ID":"8cb83be5-bf66-476c-b29e-f055352697e2","Type":"ContainerStarted","Data":"28139b9441d5c645961e3593bdeaea4281d216d56a885302e0ea045364eeb347"} Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.517730 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" event={"ID":"0a812097-b357-4b5e-bfa5-115b26efc4e3","Type":"ContainerStarted","Data":"6993a5a4ff2cdd6729f70778f32bcef5649f93e44d671c77b97896bba97c17e4"} Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.518566 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" event={"ID":"391b404f-2c22-46cf-954e-84540c6fe899","Type":"ContainerStarted","Data":"b63b2bc1608c60a74e0c5a3b56563f98a86456a062971d95282f5993c50f8e6d"} Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.520216 5110 generic.go:358] "Generic (PLEG): container finished" podID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerID="b7063645733044d69d32a09189d8bf0a789f1fbb9a6833c451c8044d0d4b5da4" exitCode=0 Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.520304 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" event={"ID":"28ec609b-6d68-482d-8ce4-2a21d490f76a","Type":"ContainerDied","Data":"b7063645733044d69d32a09189d8bf0a789f1fbb9a6833c451c8044d0d4b5da4"} Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.521396 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" event={"ID":"ccfae76d-51c9-4f8c-a694-72d770811128","Type":"ContainerStarted","Data":"2154c88fbbedc3e291ef72e94002a0f4b1c123d332c60de110eed1e68269412a"} Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.742517 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-98vxs"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.746536 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.754077 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"openshift-service-ca.crt\"" Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.757491 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"kube-root-ca.crt\"" Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.758148 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"interconnect-operator-dockercfg-2c792\"" Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.760419 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-98vxs"] Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.873706 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qv7\" (UniqueName: \"kubernetes.io/projected/17db5ab7-8c96-436c-957a-4e58f4935529-kube-api-access-l9qv7\") pod \"interconnect-operator-78b9bd8798-98vxs\" (UID: \"17db5ab7-8c96-436c-957a-4e58f4935529\") " pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" Mar 17 18:59:30 crc kubenswrapper[5110]: I0317 18:59:30.975619 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qv7\" (UniqueName: \"kubernetes.io/projected/17db5ab7-8c96-436c-957a-4e58f4935529-kube-api-access-l9qv7\") pod \"interconnect-operator-78b9bd8798-98vxs\" (UID: \"17db5ab7-8c96-436c-957a-4e58f4935529\") " pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" Mar 17 18:59:31 crc kubenswrapper[5110]: I0317 18:59:31.022364 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qv7\" (UniqueName: \"kubernetes.io/projected/17db5ab7-8c96-436c-957a-4e58f4935529-kube-api-access-l9qv7\") pod \"interconnect-operator-78b9bd8798-98vxs\" (UID: \"17db5ab7-8c96-436c-957a-4e58f4935529\") " pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" Mar 17 18:59:31 crc kubenswrapper[5110]: I0317 18:59:31.071542 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" Mar 17 18:59:31 crc kubenswrapper[5110]: I0317 18:59:31.479803 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-98vxs"] Mar 17 18:59:31 crc kubenswrapper[5110]: W0317 18:59:31.511038 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17db5ab7_8c96_436c_957a_4e58f4935529.slice/crio-ecc63d534bc3b4a0d32c5046cc793e51a0f739854466f11141befc20b4f8670f WatchSource:0}: Error finding container ecc63d534bc3b4a0d32c5046cc793e51a0f739854466f11141befc20b4f8670f: Status 404 returned error can't find the container with id ecc63d534bc3b4a0d32c5046cc793e51a0f739854466f11141befc20b4f8670f Mar 17 18:59:31 crc kubenswrapper[5110]: I0317 18:59:31.556732 5110 generic.go:358] "Generic (PLEG): container finished" podID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerID="f52723a86c305a5439ae481311ff14910a4eb92d97940ae9a0414415ee46c692" exitCode=0 Mar 17 18:59:31 crc kubenswrapper[5110]: I0317 18:59:31.556999 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" event={"ID":"28ec609b-6d68-482d-8ce4-2a21d490f76a","Type":"ContainerDied","Data":"f52723a86c305a5439ae481311ff14910a4eb92d97940ae9a0414415ee46c692"} Mar 17 18:59:31 crc kubenswrapper[5110]: I0317 18:59:31.559598 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" event={"ID":"17db5ab7-8c96-436c-957a-4e58f4935529","Type":"ContainerStarted","Data":"ecc63d534bc3b4a0d32c5046cc793e51a0f739854466f11141befc20b4f8670f"} Mar 17 18:59:32 crc kubenswrapper[5110]: I0317 18:59:32.973641 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.013896 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-bundle\") pod \"28ec609b-6d68-482d-8ce4-2a21d490f76a\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.013936 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-util\") pod \"28ec609b-6d68-482d-8ce4-2a21d490f76a\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.014009 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7lwh\" (UniqueName: \"kubernetes.io/projected/28ec609b-6d68-482d-8ce4-2a21d490f76a-kube-api-access-b7lwh\") pod \"28ec609b-6d68-482d-8ce4-2a21d490f76a\" (UID: \"28ec609b-6d68-482d-8ce4-2a21d490f76a\") " Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.019986 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-bundle" (OuterVolumeSpecName: "bundle") pod "28ec609b-6d68-482d-8ce4-2a21d490f76a" (UID: "28ec609b-6d68-482d-8ce4-2a21d490f76a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.022415 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28ec609b-6d68-482d-8ce4-2a21d490f76a-kube-api-access-b7lwh" (OuterVolumeSpecName: "kube-api-access-b7lwh") pod "28ec609b-6d68-482d-8ce4-2a21d490f76a" (UID: "28ec609b-6d68-482d-8ce4-2a21d490f76a"). InnerVolumeSpecName "kube-api-access-b7lwh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.030030 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-util" (OuterVolumeSpecName: "util") pod "28ec609b-6d68-482d-8ce4-2a21d490f76a" (UID: "28ec609b-6d68-482d-8ce4-2a21d490f76a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.115460 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-b7lwh\" (UniqueName: \"kubernetes.io/projected/28ec609b-6d68-482d-8ce4-2a21d490f76a-kube-api-access-b7lwh\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.115508 5110 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.115520 5110 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ec609b-6d68-482d-8ce4-2a21d490f76a-util\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.610191 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.610865 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br" event={"ID":"28ec609b-6d68-482d-8ce4-2a21d490f76a","Type":"ContainerDied","Data":"1552ab4b362c66e9f4431b36012e36c2241d7be22c1dc5a8343ad3839a54dd22"} Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.610902 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1552ab4b362c66e9f4431b36012e36c2241d7be22c1dc5a8343ad3839a54dd22" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.985850 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-749cd4c46b-clcvs"] Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987073 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="extract" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987090 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="extract" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987107 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="pull" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987113 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="pull" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987128 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="util" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987135 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="util" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.987286 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="28ec609b-6d68-482d-8ce4-2a21d490f76a" containerName="extract" Mar 17 18:59:33 crc kubenswrapper[5110]: I0317 18:59:33.997845 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.001516 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elastic-operator-service-cert\"" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.001704 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elastic-operator-dockercfg-gzt99\"" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.038728 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62wv8\" (UniqueName: \"kubernetes.io/projected/90161d2f-25d3-422c-8584-c8a3fd216c2e-kube-api-access-62wv8\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.038970 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/90161d2f-25d3-422c-8584-c8a3fd216c2e-webhook-cert\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.039133 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/90161d2f-25d3-422c-8584-c8a3fd216c2e-apiservice-cert\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.095469 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-749cd4c46b-clcvs"] Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.140885 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/90161d2f-25d3-422c-8584-c8a3fd216c2e-apiservice-cert\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.140947 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-62wv8\" (UniqueName: \"kubernetes.io/projected/90161d2f-25d3-422c-8584-c8a3fd216c2e-kube-api-access-62wv8\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.140998 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/90161d2f-25d3-422c-8584-c8a3fd216c2e-webhook-cert\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.152931 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/90161d2f-25d3-422c-8584-c8a3fd216c2e-webhook-cert\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.173343 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-62wv8\" (UniqueName: \"kubernetes.io/projected/90161d2f-25d3-422c-8584-c8a3fd216c2e-kube-api-access-62wv8\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.184709 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/90161d2f-25d3-422c-8584-c8a3fd216c2e-apiservice-cert\") pod \"elastic-operator-749cd4c46b-clcvs\" (UID: \"90161d2f-25d3-422c-8584-c8a3fd216c2e\") " pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.320890 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" Mar 17 18:59:34 crc kubenswrapper[5110]: I0317 18:59:34.659643 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-749cd4c46b-clcvs"] Mar 17 18:59:34 crc kubenswrapper[5110]: W0317 18:59:34.692176 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90161d2f_25d3_422c_8584_c8a3fd216c2e.slice/crio-546d919b8d0baee99eaeb5e43c809866242312d238fbede542177ef9d21b5ac4 WatchSource:0}: Error finding container 546d919b8d0baee99eaeb5e43c809866242312d238fbede542177ef9d21b5ac4: Status 404 returned error can't find the container with id 546d919b8d0baee99eaeb5e43c809866242312d238fbede542177ef9d21b5ac4 Mar 17 18:59:35 crc kubenswrapper[5110]: I0317 18:59:35.654813 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" event={"ID":"90161d2f-25d3-422c-8584-c8a3fd216c2e","Type":"ContainerStarted","Data":"546d919b8d0baee99eaeb5e43c809866242312d238fbede542177ef9d21b5ac4"} Mar 17 18:59:37 crc kubenswrapper[5110]: I0317 18:59:37.981888 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z69h6"] Mar 17 18:59:37 crc kubenswrapper[5110]: I0317 18:59:37.990493 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.007658 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z69h6"] Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.106119 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86b68\" (UniqueName: \"kubernetes.io/projected/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-kube-api-access-86b68\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.106201 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-catalog-content\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.106239 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-utilities\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.207914 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-catalog-content\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.207973 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-utilities\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.208077 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-86b68\" (UniqueName: \"kubernetes.io/projected/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-kube-api-access-86b68\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.208481 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-catalog-content\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.208567 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-utilities\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.238033 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-86b68\" (UniqueName: \"kubernetes.io/projected/53d18b2e-06d7-4ed4-bdb3-ce4766a804c1-kube-api-access-86b68\") pod \"community-operators-z69h6\" (UID: \"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1\") " pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:38 crc kubenswrapper[5110]: I0317 18:59:38.314093 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z69h6" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.801943 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z69h6"] Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.807036 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" event={"ID":"0a812097-b357-4b5e-bfa5-115b26efc4e3","Type":"ContainerStarted","Data":"acb22c663331b9d0cf9a8093b392fc44f58441f59433e29e650b736d42946af5"} Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.808034 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.809342 5110 patch_prober.go:28] interesting pod/observability-operator-587f9c8867-p6ncg container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/healthz\": dial tcp 10.217.0.47:8081: connect: connection refused" start-of-body= Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.809391 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" podUID="0a812097-b357-4b5e-bfa5-115b26efc4e3" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.47:8081/healthz\": dial tcp 10.217.0.47:8081: connect: connection refused" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.810616 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" event={"ID":"391b404f-2c22-46cf-954e-84540c6fe899","Type":"ContainerStarted","Data":"f8e54e3443fc976a12b9caacc0d9182edc48fa020f39aaa351a9de0903b81128"} Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.812880 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" event={"ID":"ccfae76d-51c9-4f8c-a694-72d770811128","Type":"ContainerStarted","Data":"e84f7ae921569da647c4473274a968bcf38ea6dc69fa79cf5abc02dc78782f96"} Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.814604 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" event={"ID":"90161d2f-25d3-422c-8584-c8a3fd216c2e","Type":"ContainerStarted","Data":"c5502279a994192413afab3daa00c2cbffe147b65130d96859a7b8823544edec"} Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.816099 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" event={"ID":"17db5ab7-8c96-436c-957a-4e58f4935529","Type":"ContainerStarted","Data":"7851ef9a2771ea86afa670e10c5a4082a48e35cab204d81d3dcef1e1d3f0e68d"} Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.817675 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5b659947cf-krgf9" event={"ID":"c9354dcb-0033-473e-a3f3-e01e7b35d9eb","Type":"ContainerStarted","Data":"9c0fd7ca6edddac65226bb5d937b5fbc2f667e93df44a06d405b35b90bfe27fe"} Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.817751 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.819287 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" event={"ID":"8cb83be5-bf66-476c-b29e-f055352697e2","Type":"ContainerStarted","Data":"6d6b0c0e78369ac704d2efbb88357ff6f1255b2c5449738fc9e66b972da00ab5"} Mar 17 18:59:48 crc kubenswrapper[5110]: W0317 18:59:48.819957 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53d18b2e_06d7_4ed4_bdb3_ce4766a804c1.slice/crio-b1aaf3a1c0213254016f923fdda7ec851f3e6356faaa0c1c1b2d86752ca89231 WatchSource:0}: Error finding container b1aaf3a1c0213254016f923fdda7ec851f3e6356faaa0c1c1b2d86752ca89231: Status 404 returned error can't find the container with id b1aaf3a1c0213254016f923fdda7ec851f3e6356faaa0c1c1b2d86752ca89231 Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.836103 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" podStartSLOduration=2.760922029 podStartE2EDuration="20.836086155s" podCreationTimestamp="2026-03-17 18:59:28 +0000 UTC" firstStartedPulling="2026-03-17 18:59:30.363556074 +0000 UTC m=+714.396167596" lastFinishedPulling="2026-03-17 18:59:48.4387202 +0000 UTC m=+732.471331722" observedRunningTime="2026-03-17 18:59:48.830786203 +0000 UTC m=+732.863397735" watchObservedRunningTime="2026-03-17 18:59:48.836086155 +0000 UTC m=+732.868697687" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.851359 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx" podStartSLOduration=3.771692157 podStartE2EDuration="21.851341143s" podCreationTimestamp="2026-03-17 18:59:27 +0000 UTC" firstStartedPulling="2026-03-17 18:59:30.350375741 +0000 UTC m=+714.382987263" lastFinishedPulling="2026-03-17 18:59:48.430024717 +0000 UTC m=+732.462636249" observedRunningTime="2026-03-17 18:59:48.845664481 +0000 UTC m=+732.878275993" watchObservedRunningTime="2026-03-17 18:59:48.851341143 +0000 UTC m=+732.883952665" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.902227 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646cb684f4-96h7l" podStartSLOduration=3.8450529209999997 podStartE2EDuration="21.902210685s" podCreationTimestamp="2026-03-17 18:59:27 +0000 UTC" firstStartedPulling="2026-03-17 18:59:30.326137032 +0000 UTC m=+714.358748564" lastFinishedPulling="2026-03-17 18:59:48.383294806 +0000 UTC m=+732.415906328" observedRunningTime="2026-03-17 18:59:48.878737366 +0000 UTC m=+732.911348888" watchObservedRunningTime="2026-03-17 18:59:48.902210685 +0000 UTC m=+732.934822207" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.902817 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5b659947cf-krgf9" podStartSLOduration=2.931108204 podStartE2EDuration="20.902810261s" podCreationTimestamp="2026-03-17 18:59:28 +0000 UTC" firstStartedPulling="2026-03-17 18:59:30.45794097 +0000 UTC m=+714.490552492" lastFinishedPulling="2026-03-17 18:59:48.429643037 +0000 UTC m=+732.462254549" observedRunningTime="2026-03-17 18:59:48.899983935 +0000 UTC m=+732.932595467" watchObservedRunningTime="2026-03-17 18:59:48.902810261 +0000 UTC m=+732.935421783" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.943809 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-749cd4c46b-clcvs" podStartSLOduration=2.276540756 podStartE2EDuration="15.943782817s" podCreationTimestamp="2026-03-17 18:59:33 +0000 UTC" firstStartedPulling="2026-03-17 18:59:34.694470227 +0000 UTC m=+718.727081749" lastFinishedPulling="2026-03-17 18:59:48.361712288 +0000 UTC m=+732.394323810" observedRunningTime="2026-03-17 18:59:48.938093855 +0000 UTC m=+732.970705387" watchObservedRunningTime="2026-03-17 18:59:48.943782817 +0000 UTC m=+732.976394359" Mar 17 18:59:48 crc kubenswrapper[5110]: I0317 18:59:48.990293 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-78b9bd8798-98vxs" podStartSLOduration=2.050824573 podStartE2EDuration="18.990271812s" podCreationTimestamp="2026-03-17 18:59:30 +0000 UTC" firstStartedPulling="2026-03-17 18:59:31.52503237 +0000 UTC m=+715.557643892" lastFinishedPulling="2026-03-17 18:59:48.464479609 +0000 UTC m=+732.497091131" observedRunningTime="2026-03-17 18:59:48.973651767 +0000 UTC m=+733.006263299" watchObservedRunningTime="2026-03-17 18:59:48.990271812 +0000 UTC m=+733.022883334" Mar 17 18:59:49 crc kubenswrapper[5110]: I0317 18:59:49.826436 5110 generic.go:358] "Generic (PLEG): container finished" podID="53d18b2e-06d7-4ed4-bdb3-ce4766a804c1" containerID="b0f046f76ed9b1a1c3553770a595b32810d713190d1baccc5300b4eddf88e2b9" exitCode=0 Mar 17 18:59:49 crc kubenswrapper[5110]: I0317 18:59:49.827015 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z69h6" event={"ID":"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1","Type":"ContainerDied","Data":"b0f046f76ed9b1a1c3553770a595b32810d713190d1baccc5300b4eddf88e2b9"} Mar 17 18:59:49 crc kubenswrapper[5110]: I0317 18:59:49.827068 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z69h6" event={"ID":"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1","Type":"ContainerStarted","Data":"b1aaf3a1c0213254016f923fdda7ec851f3e6356faaa0c1c1b2d86752ca89231"} Mar 17 18:59:49 crc kubenswrapper[5110]: I0317 18:59:49.830167 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-587f9c8867-p6ncg" Mar 17 18:59:49 crc kubenswrapper[5110]: I0317 18:59:49.861022 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-55568fc96c-6pbtm" podStartSLOduration=4.70326213 podStartE2EDuration="22.861004536s" podCreationTimestamp="2026-03-17 18:59:27 +0000 UTC" firstStartedPulling="2026-03-17 18:59:30.219810706 +0000 UTC m=+714.252422228" lastFinishedPulling="2026-03-17 18:59:48.377553112 +0000 UTC m=+732.410164634" observedRunningTime="2026-03-17 18:59:48.998520182 +0000 UTC m=+733.031131704" watchObservedRunningTime="2026-03-17 18:59:49.861004536 +0000 UTC m=+733.893616068" Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.904265 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc"] Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.913444 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.917373 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"openshift-service-ca.crt\"" Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.917789 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"kube-root-ca.crt\"" Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.921101 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager-operator\"/\"cert-manager-operator-controller-manager-dockercfg-zjprs\"" Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.937133 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc"] Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.976925 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/191f10d8-a369-48d1-a3af-1e8aeb200c8e-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pwscc\" (UID: \"191f10d8-a369-48d1-a3af-1e8aeb200c8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:50 crc kubenswrapper[5110]: I0317 18:59:50.977001 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtgjk\" (UniqueName: \"kubernetes.io/projected/191f10d8-a369-48d1-a3af-1e8aeb200c8e-kube-api-access-rtgjk\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pwscc\" (UID: \"191f10d8-a369-48d1-a3af-1e8aeb200c8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.077641 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/191f10d8-a369-48d1-a3af-1e8aeb200c8e-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pwscc\" (UID: \"191f10d8-a369-48d1-a3af-1e8aeb200c8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.077722 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rtgjk\" (UniqueName: \"kubernetes.io/projected/191f10d8-a369-48d1-a3af-1e8aeb200c8e-kube-api-access-rtgjk\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pwscc\" (UID: \"191f10d8-a369-48d1-a3af-1e8aeb200c8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.078158 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/191f10d8-a369-48d1-a3af-1e8aeb200c8e-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pwscc\" (UID: \"191f10d8-a369-48d1-a3af-1e8aeb200c8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.110015 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtgjk\" (UniqueName: \"kubernetes.io/projected/191f10d8-a369-48d1-a3af-1e8aeb200c8e-kube-api-access-rtgjk\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pwscc\" (UID: \"191f10d8-a369-48d1-a3af-1e8aeb200c8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.230211 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.493830 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc"] Mar 17 18:59:51 crc kubenswrapper[5110]: W0317 18:59:51.502950 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod191f10d8_a369_48d1_a3af_1e8aeb200c8e.slice/crio-35027d26e52d68ec873f2bfd1a8dc946136445262987e4022ef8c151db64cb49 WatchSource:0}: Error finding container 35027d26e52d68ec873f2bfd1a8dc946136445262987e4022ef8c151db64cb49: Status 404 returned error can't find the container with id 35027d26e52d68ec873f2bfd1a8dc946136445262987e4022ef8c151db64cb49 Mar 17 18:59:51 crc kubenswrapper[5110]: I0317 18:59:51.846952 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" event={"ID":"191f10d8-a369-48d1-a3af-1e8aeb200c8e","Type":"ContainerStarted","Data":"35027d26e52d68ec873f2bfd1a8dc946136445262987e4022ef8c151db64cb49"} Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.430850 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" podUID="67450a68-b2ce-4acb-ba06-f4e60fcf28b4" containerName="registry" containerID="cri-o://75579bcc7f6b04e5e922a381669b2715f243ef1666a95d2bded26a21b8a6a387" gracePeriod=30 Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.861661 5110 generic.go:358] "Generic (PLEG): container finished" podID="67450a68-b2ce-4acb-ba06-f4e60fcf28b4" containerID="75579bcc7f6b04e5e922a381669b2715f243ef1666a95d2bded26a21b8a6a387" exitCode=0 Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.861715 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" event={"ID":"67450a68-b2ce-4acb-ba06-f4e60fcf28b4","Type":"ContainerDied","Data":"75579bcc7f6b04e5e922a381669b2715f243ef1666a95d2bded26a21b8a6a387"} Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.861771 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" event={"ID":"67450a68-b2ce-4acb-ba06-f4e60fcf28b4","Type":"ContainerDied","Data":"c32fda45b100eedbdc0125f07d0e7e348e200caccee6a6ffc4c0f35c94322e2a"} Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.861784 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c32fda45b100eedbdc0125f07d0e7e348e200caccee6a6ffc4c0f35c94322e2a" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.884586 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.904225 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.904309 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-bound-sa-token\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.904359 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nwzw\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-kube-api-access-4nwzw\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.904390 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-tls\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.904408 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-certificates\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.904444 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-trusted-ca\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.905136 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.905687 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.907031 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-installation-pull-secrets\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.907112 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-ca-trust-extracted\") pod \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\" (UID: \"67450a68-b2ce-4acb-ba06-f4e60fcf28b4\") " Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.907932 5110 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.907949 5110 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.933255 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-kube-api-access-4nwzw" (OuterVolumeSpecName: "kube-api-access-4nwzw") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "kube-api-access-4nwzw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.937194 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.939202 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.944644 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.953575 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 18:59:52 crc kubenswrapper[5110]: I0317 18:59:52.989957 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (OuterVolumeSpecName: "registry-storage") pod "67450a68-b2ce-4acb-ba06-f4e60fcf28b4" (UID: "67450a68-b2ce-4acb-ba06-f4e60fcf28b4"). InnerVolumeSpecName "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2". PluginName "kubernetes.io/csi", VolumeGIDValue "" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.009286 5110 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.009326 5110 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.009340 5110 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.009352 5110 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.009364 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4nwzw\" (UniqueName: \"kubernetes.io/projected/67450a68-b2ce-4acb-ba06-f4e60fcf28b4-kube-api-access-4nwzw\") on node \"crc\" DevicePath \"\"" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.872433 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-5hkdz" Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.919136 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-5hkdz"] Mar 17 18:59:53 crc kubenswrapper[5110]: I0317 18:59:53.954424 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-5hkdz"] Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.754902 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67450a68-b2ce-4acb-ba06-f4e60fcf28b4" path="/var/lib/kubelet/pods/67450a68-b2ce-4acb-ba06-f4e60fcf28b4/volumes" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.860433 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.860993 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="67450a68-b2ce-4acb-ba06-f4e60fcf28b4" containerName="registry" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.861007 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="67450a68-b2ce-4acb-ba06-f4e60fcf28b4" containerName="registry" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.861140 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="67450a68-b2ce-4acb-ba06-f4e60fcf28b4" containerName="registry" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.869813 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.871826 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-remote-ca\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.872688 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-http-certs-internal\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.872967 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-default-es-config\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.873832 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-default-es-transport-certs\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.874151 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"elasticsearch-es-unicast-hosts\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.874337 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-dockercfg-pq62r\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.874554 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-internal-users\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.874572 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-xpack-file-realm\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.875126 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"elasticsearch-es-scripts\"" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.884469 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.939886 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940019 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940128 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940151 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940169 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940202 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940231 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940271 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940295 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940316 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940350 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/d9344573-098c-4529-a15f-b84912c0b147-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940406 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940432 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940558 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:54 crc kubenswrapper[5110]: I0317 18:59:54.940596 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041655 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/d9344573-098c-4529-a15f-b84912c0b147-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041732 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041758 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041793 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041814 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041855 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041875 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041916 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041938 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041960 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.041989 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.042018 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.042065 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.042092 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.042117 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.043075 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.043088 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.043613 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.043695 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.043958 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.046377 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.047451 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.050653 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.052270 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/d9344573-098c-4529-a15f-b84912c0b147-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.055733 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.056168 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.056660 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.066941 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.076107 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/d9344573-098c-4529-a15f-b84912c0b147-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.076354 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/d9344573-098c-4529-a15f-b84912c0b147-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"d9344573-098c-4529-a15f-b84912c0b147\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:55 crc kubenswrapper[5110]: I0317 18:59:55.187916 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 17 18:59:58 crc kubenswrapper[5110]: I0317 18:59:58.043418 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 17 18:59:58 crc kubenswrapper[5110]: I0317 18:59:58.910310 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" event={"ID":"191f10d8-a369-48d1-a3af-1e8aeb200c8e","Type":"ContainerStarted","Data":"f075e993323f38638d60bc08565b4ecc584a3f8e9ac208f80299a5bf3fc88014"} Mar 17 18:59:58 crc kubenswrapper[5110]: I0317 18:59:58.927295 5110 generic.go:358] "Generic (PLEG): container finished" podID="53d18b2e-06d7-4ed4-bdb3-ce4766a804c1" containerID="4b71859307bb7f2e7d7831f5616dc4490a659e1f206f0dcc909305bd7a7d40b7" exitCode=0 Mar 17 18:59:58 crc kubenswrapper[5110]: I0317 18:59:58.927775 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z69h6" event={"ID":"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1","Type":"ContainerDied","Data":"4b71859307bb7f2e7d7831f5616dc4490a659e1f206f0dcc909305bd7a7d40b7"} Mar 17 18:59:58 crc kubenswrapper[5110]: I0317 18:59:58.933269 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"d9344573-098c-4529-a15f-b84912c0b147","Type":"ContainerStarted","Data":"f2b882c0851fe83e9dbe143c86177d47b8ef92441521188e342a915670345f7e"} Mar 17 18:59:58 crc kubenswrapper[5110]: I0317 18:59:58.963047 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pwscc" podStartSLOduration=2.628679329 podStartE2EDuration="8.963031943s" podCreationTimestamp="2026-03-17 18:59:50 +0000 UTC" firstStartedPulling="2026-03-17 18:59:51.505690564 +0000 UTC m=+735.538302086" lastFinishedPulling="2026-03-17 18:59:57.840043178 +0000 UTC m=+741.872654700" observedRunningTime="2026-03-17 18:59:58.933446121 +0000 UTC m=+742.966057643" watchObservedRunningTime="2026-03-17 18:59:58.963031943 +0000 UTC m=+742.995643455" Mar 17 18:59:59 crc kubenswrapper[5110]: I0317 18:59:59.831891 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5b659947cf-krgf9" Mar 17 18:59:59 crc kubenswrapper[5110]: I0317 18:59:59.952827 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z69h6" event={"ID":"53d18b2e-06d7-4ed4-bdb3-ce4766a804c1","Type":"ContainerStarted","Data":"bc06394530b984917dea296f8265d1eedc36fe3f8f60c5cd59848009d9f2c89c"} Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.000179 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z69h6" podStartSLOduration=14.963862077 podStartE2EDuration="23.000159151s" podCreationTimestamp="2026-03-17 18:59:37 +0000 UTC" firstStartedPulling="2026-03-17 18:59:49.827247293 +0000 UTC m=+733.859858815" lastFinishedPulling="2026-03-17 18:59:57.863544367 +0000 UTC m=+741.896155889" observedRunningTime="2026-03-17 18:59:59.995693842 +0000 UTC m=+744.028305364" watchObservedRunningTime="2026-03-17 19:00:00.000159151 +0000 UTC m=+744.032770673" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.136977 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp"] Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.147602 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.150156 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.150730 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.153405 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp"] Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.214745 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114ed064-369c-4729-9957-3333e8288e90-config-volume\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.214789 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/114ed064-369c-4729-9957-3333e8288e90-secret-volume\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.214872 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sscxp\" (UniqueName: \"kubernetes.io/projected/114ed064-369c-4729-9957-3333e8288e90-kube-api-access-sscxp\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.230209 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562900-hsptz"] Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.234649 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.237726 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.237976 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.238300 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.241749 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562900-hsptz"] Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.315723 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9slsg\" (UniqueName: \"kubernetes.io/projected/7ca21359-d4f9-4720-8e0f-21c5b449d62f-kube-api-access-9slsg\") pod \"auto-csr-approver-29562900-hsptz\" (UID: \"7ca21359-d4f9-4720-8e0f-21c5b449d62f\") " pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.315807 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sscxp\" (UniqueName: \"kubernetes.io/projected/114ed064-369c-4729-9957-3333e8288e90-kube-api-access-sscxp\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.315864 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114ed064-369c-4729-9957-3333e8288e90-config-volume\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.315886 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/114ed064-369c-4729-9957-3333e8288e90-secret-volume\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.318095 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114ed064-369c-4729-9957-3333e8288e90-config-volume\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.323295 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/114ed064-369c-4729-9957-3333e8288e90-secret-volume\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.349116 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sscxp\" (UniqueName: \"kubernetes.io/projected/114ed064-369c-4729-9957-3333e8288e90-kube-api-access-sscxp\") pod \"collect-profiles-29562900-qnrhp\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.417065 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9slsg\" (UniqueName: \"kubernetes.io/projected/7ca21359-d4f9-4720-8e0f-21c5b449d62f-kube-api-access-9slsg\") pod \"auto-csr-approver-29562900-hsptz\" (UID: \"7ca21359-d4f9-4720-8e0f-21c5b449d62f\") " pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.451907 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9slsg\" (UniqueName: \"kubernetes.io/projected/7ca21359-d4f9-4720-8e0f-21c5b449d62f-kube-api-access-9slsg\") pod \"auto-csr-approver-29562900-hsptz\" (UID: \"7ca21359-d4f9-4720-8e0f-21c5b449d62f\") " pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.477544 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.551856 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.944395 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562900-hsptz"] Mar 17 19:00:00 crc kubenswrapper[5110]: I0317 19:00:00.985124 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562900-hsptz" event={"ID":"7ca21359-d4f9-4720-8e0f-21c5b449d62f","Type":"ContainerStarted","Data":"de99fe20d5169c94a7ea7b77f835d108da8e0a1697392a1d223cb6a23706c03e"} Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.039802 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp"] Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.313495 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-nzg8p"] Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.317617 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.319718 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.323153 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-jz6pp\"" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.323165 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.327325 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-nzg8p"] Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.447332 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdpv6\" (UniqueName: \"kubernetes.io/projected/88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1-kube-api-access-xdpv6\") pod \"cert-manager-webhook-597b96b99b-nzg8p\" (UID: \"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1\") " pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.447411 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-nzg8p\" (UID: \"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1\") " pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.548675 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-nzg8p\" (UID: \"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1\") " pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.548791 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xdpv6\" (UniqueName: \"kubernetes.io/projected/88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1-kube-api-access-xdpv6\") pod \"cert-manager-webhook-597b96b99b-nzg8p\" (UID: \"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1\") " pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.581990 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-nzg8p\" (UID: \"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1\") " pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.585848 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdpv6\" (UniqueName: \"kubernetes.io/projected/88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1-kube-api-access-xdpv6\") pod \"cert-manager-webhook-597b96b99b-nzg8p\" (UID: \"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1\") " pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.676021 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.995094 5110 generic.go:358] "Generic (PLEG): container finished" podID="114ed064-369c-4729-9957-3333e8288e90" containerID="cc2b7307c2034a8ea2fe0615e5a4373b092ccc21862029f6accd1f733a8fd02b" exitCode=0 Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.995175 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" event={"ID":"114ed064-369c-4729-9957-3333e8288e90","Type":"ContainerDied","Data":"cc2b7307c2034a8ea2fe0615e5a4373b092ccc21862029f6accd1f733a8fd02b"} Mar 17 19:00:01 crc kubenswrapper[5110]: I0317 19:00:01.995200 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" event={"ID":"114ed064-369c-4729-9957-3333e8288e90","Type":"ContainerStarted","Data":"0ad3863d72cc236e4342fa488350f469aeeb12b945550ec43224066d0b0c5d93"} Mar 17 19:00:02 crc kubenswrapper[5110]: I0317 19:00:02.195925 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-nzg8p"] Mar 17 19:00:02 crc kubenswrapper[5110]: W0317 19:00:02.206986 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88dd76ec_ef69_4d92_aaaf_bbe85e63b5c1.slice/crio-8af39ff5bc567d2e091c0d759ab2175526ee6ac1569cfa16f7b4e36f52b91d9b WatchSource:0}: Error finding container 8af39ff5bc567d2e091c0d759ab2175526ee6ac1569cfa16f7b4e36f52b91d9b: Status 404 returned error can't find the container with id 8af39ff5bc567d2e091c0d759ab2175526ee6ac1569cfa16f7b4e36f52b91d9b Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.009450 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" event={"ID":"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1","Type":"ContainerStarted","Data":"8af39ff5bc567d2e091c0d759ab2175526ee6ac1569cfa16f7b4e36f52b91d9b"} Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.320823 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.479633 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sscxp\" (UniqueName: \"kubernetes.io/projected/114ed064-369c-4729-9957-3333e8288e90-kube-api-access-sscxp\") pod \"114ed064-369c-4729-9957-3333e8288e90\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.479694 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114ed064-369c-4729-9957-3333e8288e90-config-volume\") pod \"114ed064-369c-4729-9957-3333e8288e90\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.479750 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/114ed064-369c-4729-9957-3333e8288e90-secret-volume\") pod \"114ed064-369c-4729-9957-3333e8288e90\" (UID: \"114ed064-369c-4729-9957-3333e8288e90\") " Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.482423 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/114ed064-369c-4729-9957-3333e8288e90-config-volume" (OuterVolumeSpecName: "config-volume") pod "114ed064-369c-4729-9957-3333e8288e90" (UID: "114ed064-369c-4729-9957-3333e8288e90"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.487304 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/114ed064-369c-4729-9957-3333e8288e90-kube-api-access-sscxp" (OuterVolumeSpecName: "kube-api-access-sscxp") pod "114ed064-369c-4729-9957-3333e8288e90" (UID: "114ed064-369c-4729-9957-3333e8288e90"). InnerVolumeSpecName "kube-api-access-sscxp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.488199 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/114ed064-369c-4729-9957-3333e8288e90-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "114ed064-369c-4729-9957-3333e8288e90" (UID: "114ed064-369c-4729-9957-3333e8288e90"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.581228 5110 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/114ed064-369c-4729-9957-3333e8288e90-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.581269 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sscxp\" (UniqueName: \"kubernetes.io/projected/114ed064-369c-4729-9957-3333e8288e90-kube-api-access-sscxp\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:03 crc kubenswrapper[5110]: I0317 19:00:03.581282 5110 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/114ed064-369c-4729-9957-3333e8288e90-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:04 crc kubenswrapper[5110]: I0317 19:00:04.016094 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" event={"ID":"114ed064-369c-4729-9957-3333e8288e90","Type":"ContainerDied","Data":"0ad3863d72cc236e4342fa488350f469aeeb12b945550ec43224066d0b0c5d93"} Mar 17 19:00:04 crc kubenswrapper[5110]: I0317 19:00:04.016131 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562900-qnrhp" Mar 17 19:00:04 crc kubenswrapper[5110]: I0317 19:00:04.016137 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ad3863d72cc236e4342fa488350f469aeeb12b945550ec43224066d0b0c5d93" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.042590 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-5mcvs"] Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.043277 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="114ed064-369c-4729-9957-3333e8288e90" containerName="collect-profiles" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.043294 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="114ed064-369c-4729-9957-3333e8288e90" containerName="collect-profiles" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.043419 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="114ed064-369c-4729-9957-3333e8288e90" containerName="collect-profiles" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.080400 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-5mcvs"] Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.080537 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.083148 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-6bvgh\"" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.203900 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0495d1f8-565a-45e1-ad95-c595cf5a894c-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-5mcvs\" (UID: \"0495d1f8-565a-45e1-ad95-c595cf5a894c\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.204207 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8qvs\" (UniqueName: \"kubernetes.io/projected/0495d1f8-565a-45e1-ad95-c595cf5a894c-kube-api-access-v8qvs\") pod \"cert-manager-cainjector-8966b78d4-5mcvs\" (UID: \"0495d1f8-565a-45e1-ad95-c595cf5a894c\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.304916 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0495d1f8-565a-45e1-ad95-c595cf5a894c-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-5mcvs\" (UID: \"0495d1f8-565a-45e1-ad95-c595cf5a894c\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.304982 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v8qvs\" (UniqueName: \"kubernetes.io/projected/0495d1f8-565a-45e1-ad95-c595cf5a894c-kube-api-access-v8qvs\") pod \"cert-manager-cainjector-8966b78d4-5mcvs\" (UID: \"0495d1f8-565a-45e1-ad95-c595cf5a894c\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.332336 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0495d1f8-565a-45e1-ad95-c595cf5a894c-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-5mcvs\" (UID: \"0495d1f8-565a-45e1-ad95-c595cf5a894c\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.333031 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8qvs\" (UniqueName: \"kubernetes.io/projected/0495d1f8-565a-45e1-ad95-c595cf5a894c-kube-api-access-v8qvs\") pod \"cert-manager-cainjector-8966b78d4-5mcvs\" (UID: \"0495d1f8-565a-45e1-ad95-c595cf5a894c\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.394526 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" Mar 17 19:00:05 crc kubenswrapper[5110]: I0317 19:00:05.862930 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-5mcvs"] Mar 17 19:00:06 crc kubenswrapper[5110]: I0317 19:00:06.036333 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" event={"ID":"0495d1f8-565a-45e1-ad95-c595cf5a894c","Type":"ContainerStarted","Data":"91a1416c91f7f95ab4555f335bbd01da5b9e2bb6f0027e818293e8e8c6177bf9"} Mar 17 19:00:08 crc kubenswrapper[5110]: I0317 19:00:08.314244 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-z69h6" Mar 17 19:00:08 crc kubenswrapper[5110]: I0317 19:00:08.314589 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z69h6" Mar 17 19:00:08 crc kubenswrapper[5110]: I0317 19:00:08.356744 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z69h6" Mar 17 19:00:09 crc kubenswrapper[5110]: I0317 19:00:09.092182 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z69h6" Mar 17 19:00:09 crc kubenswrapper[5110]: I0317 19:00:09.176958 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z69h6"] Mar 17 19:00:09 crc kubenswrapper[5110]: I0317 19:00:09.214180 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t72j2"] Mar 17 19:00:09 crc kubenswrapper[5110]: I0317 19:00:09.214457 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t72j2" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="registry-server" containerID="cri-o://7c9d9e3b1f7c677c263a5bff83512434f4bfd20e3ca5de0f68576f2c291fd995" gracePeriod=2 Mar 17 19:00:10 crc kubenswrapper[5110]: I0317 19:00:10.063344 5110 generic.go:358] "Generic (PLEG): container finished" podID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerID="7c9d9e3b1f7c677c263a5bff83512434f4bfd20e3ca5de0f68576f2c291fd995" exitCode=0 Mar 17 19:00:10 crc kubenswrapper[5110]: I0317 19:00:10.063372 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerDied","Data":"7c9d9e3b1f7c677c263a5bff83512434f4bfd20e3ca5de0f68576f2c291fd995"} Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.308559 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-759f64656b-s8xn7"] Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.317849 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.320490 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-dockercfg-mn58l\"" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.321737 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-759f64656b-s8xn7"] Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.491858 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7lsj\" (UniqueName: \"kubernetes.io/projected/366966b1-9c92-430c-bc89-5a84ac8ed770-kube-api-access-l7lsj\") pod \"cert-manager-759f64656b-s8xn7\" (UID: \"366966b1-9c92-430c-bc89-5a84ac8ed770\") " pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.491930 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/366966b1-9c92-430c-bc89-5a84ac8ed770-bound-sa-token\") pod \"cert-manager-759f64656b-s8xn7\" (UID: \"366966b1-9c92-430c-bc89-5a84ac8ed770\") " pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.593708 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7lsj\" (UniqueName: \"kubernetes.io/projected/366966b1-9c92-430c-bc89-5a84ac8ed770-kube-api-access-l7lsj\") pod \"cert-manager-759f64656b-s8xn7\" (UID: \"366966b1-9c92-430c-bc89-5a84ac8ed770\") " pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.593780 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/366966b1-9c92-430c-bc89-5a84ac8ed770-bound-sa-token\") pod \"cert-manager-759f64656b-s8xn7\" (UID: \"366966b1-9c92-430c-bc89-5a84ac8ed770\") " pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.615794 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/366966b1-9c92-430c-bc89-5a84ac8ed770-bound-sa-token\") pod \"cert-manager-759f64656b-s8xn7\" (UID: \"366966b1-9c92-430c-bc89-5a84ac8ed770\") " pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.624781 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7lsj\" (UniqueName: \"kubernetes.io/projected/366966b1-9c92-430c-bc89-5a84ac8ed770-kube-api-access-l7lsj\") pod \"cert-manager-759f64656b-s8xn7\" (UID: \"366966b1-9c92-430c-bc89-5a84ac8ed770\") " pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:11 crc kubenswrapper[5110]: I0317 19:00:11.642891 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-759f64656b-s8xn7" Mar 17 19:00:12 crc kubenswrapper[5110]: I0317 19:00:12.132856 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:00:12 crc kubenswrapper[5110]: I0317 19:00:12.132918 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:00:14 crc kubenswrapper[5110]: I0317 19:00:14.843492 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t72j2" Mar 17 19:00:14 crc kubenswrapper[5110]: I0317 19:00:14.940083 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-catalog-content\") pod \"deba3474-2676-42bb-8a3a-5d47a06caf16\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " Mar 17 19:00:14 crc kubenswrapper[5110]: I0317 19:00:14.940386 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llnxd\" (UniqueName: \"kubernetes.io/projected/deba3474-2676-42bb-8a3a-5d47a06caf16-kube-api-access-llnxd\") pod \"deba3474-2676-42bb-8a3a-5d47a06caf16\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " Mar 17 19:00:14 crc kubenswrapper[5110]: I0317 19:00:14.940417 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-utilities\") pod \"deba3474-2676-42bb-8a3a-5d47a06caf16\" (UID: \"deba3474-2676-42bb-8a3a-5d47a06caf16\") " Mar 17 19:00:14 crc kubenswrapper[5110]: I0317 19:00:14.941925 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-utilities" (OuterVolumeSpecName: "utilities") pod "deba3474-2676-42bb-8a3a-5d47a06caf16" (UID: "deba3474-2676-42bb-8a3a-5d47a06caf16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:00:14 crc kubenswrapper[5110]: I0317 19:00:14.948967 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deba3474-2676-42bb-8a3a-5d47a06caf16-kube-api-access-llnxd" (OuterVolumeSpecName: "kube-api-access-llnxd") pod "deba3474-2676-42bb-8a3a-5d47a06caf16" (UID: "deba3474-2676-42bb-8a3a-5d47a06caf16"). InnerVolumeSpecName "kube-api-access-llnxd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.025181 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "deba3474-2676-42bb-8a3a-5d47a06caf16" (UID: "deba3474-2676-42bb-8a3a-5d47a06caf16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.041939 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.041975 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-llnxd\" (UniqueName: \"kubernetes.io/projected/deba3474-2676-42bb-8a3a-5d47a06caf16-kube-api-access-llnxd\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.041987 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deba3474-2676-42bb-8a3a-5d47a06caf16-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.045480 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-759f64656b-s8xn7"] Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.097353 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t72j2" event={"ID":"deba3474-2676-42bb-8a3a-5d47a06caf16","Type":"ContainerDied","Data":"74e27a1bf21c1a283e13031e84124a6632570f8192afdf2bc205f5f90cc42529"} Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.097413 5110 scope.go:117] "RemoveContainer" containerID="7c9d9e3b1f7c677c263a5bff83512434f4bfd20e3ca5de0f68576f2c291fd995" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.097554 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t72j2" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.102275 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-759f64656b-s8xn7" event={"ID":"366966b1-9c92-430c-bc89-5a84ac8ed770","Type":"ContainerStarted","Data":"61f5526daf6036fc9e71684c2a56d859f549087ae918ce5cc7d1849902347da1"} Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.123595 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t72j2"] Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.128293 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t72j2"] Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.154824 5110 scope.go:117] "RemoveContainer" containerID="39128367db3304d0be722eac652f6b3db3e7144568325a3ca0cef638f0905454" Mar 17 19:00:15 crc kubenswrapper[5110]: I0317 19:00:15.270258 5110 scope.go:117] "RemoveContainer" containerID="d4c0a61b0a353897a51ba364aa8442cca024eb2c6e0f6711eb50652942ec4e27" Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.122221 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" event={"ID":"0495d1f8-565a-45e1-ad95-c595cf5a894c","Type":"ContainerStarted","Data":"651bb84f8c31d4f9454f4af00e78b0a5dc06699afb35a351e6456acb4a65c217"} Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.124832 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"d9344573-098c-4529-a15f-b84912c0b147","Type":"ContainerStarted","Data":"cf20d2506acb298cdfe535fd7ab7af7c45c61697bda6c44449cf684af589ab4c"} Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.126545 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" event={"ID":"88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1","Type":"ContainerStarted","Data":"0a1c0b2898562c5e913b82fbb910a48d94e66f01fdfb79bef29e720e0012659a"} Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.126778 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.128048 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-759f64656b-s8xn7" event={"ID":"366966b1-9c92-430c-bc89-5a84ac8ed770","Type":"ContainerStarted","Data":"d459c04579945df31888a4f9814643715640393f2f043dbd77a107b7a6131f04"} Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.129443 5110 generic.go:358] "Generic (PLEG): container finished" podID="7ca21359-d4f9-4720-8e0f-21c5b449d62f" containerID="67391f7570881da2e65abbeed016414d864b17053df32bf0a3cf7a5fcfde1932" exitCode=0 Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.129484 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562900-hsptz" event={"ID":"7ca21359-d4f9-4720-8e0f-21c5b449d62f","Type":"ContainerDied","Data":"67391f7570881da2e65abbeed016414d864b17053df32bf0a3cf7a5fcfde1932"} Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.137503 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-8966b78d4-5mcvs" podStartSLOduration=1.938716506 podStartE2EDuration="11.137483524s" podCreationTimestamp="2026-03-17 19:00:05 +0000 UTC" firstStartedPulling="2026-03-17 19:00:05.882667851 +0000 UTC m=+749.915279373" lastFinishedPulling="2026-03-17 19:00:15.081434869 +0000 UTC m=+759.114046391" observedRunningTime="2026-03-17 19:00:16.133815475 +0000 UTC m=+760.166427027" watchObservedRunningTime="2026-03-17 19:00:16.137483524 +0000 UTC m=+760.170095046" Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.211727 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" podStartSLOduration=2.150949918 podStartE2EDuration="15.211706499s" podCreationTimestamp="2026-03-17 19:00:01 +0000 UTC" firstStartedPulling="2026-03-17 19:00:02.21055034 +0000 UTC m=+746.243161872" lastFinishedPulling="2026-03-17 19:00:15.271306931 +0000 UTC m=+759.303918453" observedRunningTime="2026-03-17 19:00:16.211655228 +0000 UTC m=+760.244266750" watchObservedRunningTime="2026-03-17 19:00:16.211706499 +0000 UTC m=+760.244318021" Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.228399 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-759f64656b-s8xn7" podStartSLOduration=5.228374755 podStartE2EDuration="5.228374755s" podCreationTimestamp="2026-03-17 19:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:00:16.223209627 +0000 UTC m=+760.255821159" watchObservedRunningTime="2026-03-17 19:00:16.228374755 +0000 UTC m=+760.260986277" Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.259027 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.286701 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 17 19:00:16 crc kubenswrapper[5110]: I0317 19:00:16.752381 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" path="/var/lib/kubelet/pods/deba3474-2676-42bb-8a3a-5d47a06caf16/volumes" Mar 17 19:00:17 crc kubenswrapper[5110]: I0317 19:00:17.417419 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:17 crc kubenswrapper[5110]: I0317 19:00:17.475299 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9slsg\" (UniqueName: \"kubernetes.io/projected/7ca21359-d4f9-4720-8e0f-21c5b449d62f-kube-api-access-9slsg\") pod \"7ca21359-d4f9-4720-8e0f-21c5b449d62f\" (UID: \"7ca21359-d4f9-4720-8e0f-21c5b449d62f\") " Mar 17 19:00:17 crc kubenswrapper[5110]: I0317 19:00:17.481665 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ca21359-d4f9-4720-8e0f-21c5b449d62f-kube-api-access-9slsg" (OuterVolumeSpecName: "kube-api-access-9slsg") pod "7ca21359-d4f9-4720-8e0f-21c5b449d62f" (UID: "7ca21359-d4f9-4720-8e0f-21c5b449d62f"). InnerVolumeSpecName "kube-api-access-9slsg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:00:17 crc kubenswrapper[5110]: I0317 19:00:17.576731 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9slsg\" (UniqueName: \"kubernetes.io/projected/7ca21359-d4f9-4720-8e0f-21c5b449d62f-kube-api-access-9slsg\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.145355 5110 generic.go:358] "Generic (PLEG): container finished" podID="d9344573-098c-4529-a15f-b84912c0b147" containerID="cf20d2506acb298cdfe535fd7ab7af7c45c61697bda6c44449cf684af589ab4c" exitCode=0 Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.145452 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"d9344573-098c-4529-a15f-b84912c0b147","Type":"ContainerDied","Data":"cf20d2506acb298cdfe535fd7ab7af7c45c61697bda6c44449cf684af589ab4c"} Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.146998 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562900-hsptz" event={"ID":"7ca21359-d4f9-4720-8e0f-21c5b449d62f","Type":"ContainerDied","Data":"de99fe20d5169c94a7ea7b77f835d108da8e0a1697392a1d223cb6a23706c03e"} Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.147040 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de99fe20d5169c94a7ea7b77f835d108da8e0a1697392a1d223cb6a23706c03e" Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.147135 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562900-hsptz" Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.472587 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562894-qrq5w"] Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.478456 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562894-qrq5w"] Mar 17 19:00:18 crc kubenswrapper[5110]: I0317 19:00:18.753500 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dc544ed-7904-47b8-9980-10d7015526ac" path="/var/lib/kubelet/pods/1dc544ed-7904-47b8-9980-10d7015526ac/volumes" Mar 17 19:00:19 crc kubenswrapper[5110]: I0317 19:00:19.156406 5110 generic.go:358] "Generic (PLEG): container finished" podID="d9344573-098c-4529-a15f-b84912c0b147" containerID="543a0718b91450b3247a0266b2b906c21d0df29c9470533a9cb4885fe6bedb74" exitCode=0 Mar 17 19:00:19 crc kubenswrapper[5110]: I0317 19:00:19.156539 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"d9344573-098c-4529-a15f-b84912c0b147","Type":"ContainerDied","Data":"543a0718b91450b3247a0266b2b906c21d0df29c9470533a9cb4885fe6bedb74"} Mar 17 19:00:20 crc kubenswrapper[5110]: I0317 19:00:20.166970 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"d9344573-098c-4529-a15f-b84912c0b147","Type":"ContainerStarted","Data":"b6d07cf50d572724c81104393ce84293d9c32b81f1e602b56e52801bf5688add"} Mar 17 19:00:20 crc kubenswrapper[5110]: I0317 19:00:20.167361 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 17 19:00:20 crc kubenswrapper[5110]: I0317 19:00:20.209357 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=8.908761007 podStartE2EDuration="26.209338612s" podCreationTimestamp="2026-03-17 18:59:54 +0000 UTC" firstStartedPulling="2026-03-17 18:59:58.058030942 +0000 UTC m=+742.090642464" lastFinishedPulling="2026-03-17 19:00:15.358608547 +0000 UTC m=+759.391220069" observedRunningTime="2026-03-17 19:00:20.207872423 +0000 UTC m=+764.240483945" watchObservedRunningTime="2026-03-17 19:00:20.209338612 +0000 UTC m=+764.241950134" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.142088 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-597b96b99b-nzg8p" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.727364 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728005 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="extract-content" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728023 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="extract-content" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728047 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7ca21359-d4f9-4720-8e0f-21c5b449d62f" containerName="oc" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728084 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca21359-d4f9-4720-8e0f-21c5b449d62f" containerName="oc" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728093 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="extract-utilities" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728099 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="extract-utilities" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728112 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="registry-server" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728118 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="registry-server" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728222 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="7ca21359-d4f9-4720-8e0f-21c5b449d62f" containerName="oc" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.728230 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="deba3474-2676-42bb-8a3a-5d47a06caf16" containerName="registry-server" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.737041 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.738971 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.739574 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-1-sys-config\"" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.739951 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-1-global-ca\"" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.740000 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-1-ca\"" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.743787 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.743865 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjrm\" (UniqueName: \"kubernetes.io/projected/601ac852-e73a-4c50-a9d6-bbc199c69460-kube-api-access-vkjrm\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.743945 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.743975 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744029 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744072 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744100 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744131 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744195 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744237 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744339 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-push\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.744384 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.758810 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.845788 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.845865 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.845895 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-push\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846016 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846038 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846141 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjrm\" (UniqueName: \"kubernetes.io/projected/601ac852-e73a-4c50-a9d6-bbc199c69460-kube-api-access-vkjrm\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846173 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846199 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846229 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846254 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846280 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846312 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846430 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846515 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846676 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.846897 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.847263 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.848112 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.848116 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.848158 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.849688 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.855956 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-push\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.865674 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:22 crc kubenswrapper[5110]: I0317 19:00:22.875331 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjrm\" (UniqueName: \"kubernetes.io/projected/601ac852-e73a-4c50-a9d6-bbc199c69460-kube-api-access-vkjrm\") pod \"service-telemetry-operator-1-build\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:23 crc kubenswrapper[5110]: I0317 19:00:23.053762 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:23 crc kubenswrapper[5110]: I0317 19:00:23.482878 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 17 19:00:24 crc kubenswrapper[5110]: I0317 19:00:24.204576 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"601ac852-e73a-4c50-a9d6-bbc199c69460","Type":"ContainerStarted","Data":"fafd90ed2609398c9893e648acf918b3913824573295e9157d770cb2e2c145ba"} Mar 17 19:00:30 crc kubenswrapper[5110]: I0317 19:00:30.241269 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"601ac852-e73a-4c50-a9d6-bbc199c69460","Type":"ContainerStarted","Data":"2ece55a9c24fb66ddfda9418f3219cc2fdeefa2356e712ca667b672c9ea0e792"} Mar 17 19:00:31 crc kubenswrapper[5110]: I0317 19:00:31.247475 5110 prober.go:120] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="d9344573-098c-4529-a15f-b84912c0b147" containerName="elasticsearch" probeResult="failure" output=< Mar 17 19:00:31 crc kubenswrapper[5110]: {"timestamp": "2026-03-17T19:00:31+00:00", "message": "readiness probe failed", "curl_rc": "7"} Mar 17 19:00:31 crc kubenswrapper[5110]: > Mar 17 19:00:31 crc kubenswrapper[5110]: I0317 19:00:31.249349 5110 generic.go:358] "Generic (PLEG): container finished" podID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerID="2ece55a9c24fb66ddfda9418f3219cc2fdeefa2356e712ca667b672c9ea0e792" exitCode=0 Mar 17 19:00:31 crc kubenswrapper[5110]: I0317 19:00:31.249414 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"601ac852-e73a-4c50-a9d6-bbc199c69460","Type":"ContainerDied","Data":"2ece55a9c24fb66ddfda9418f3219cc2fdeefa2356e712ca667b672c9ea0e792"} Mar 17 19:00:32 crc kubenswrapper[5110]: I0317 19:00:32.260022 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"601ac852-e73a-4c50-a9d6-bbc199c69460","Type":"ContainerStarted","Data":"0fb1a78f96260c1c2835b5cb665439730c0368e7d423f678de3943d86c975acf"} Mar 17 19:00:32 crc kubenswrapper[5110]: I0317 19:00:32.291767 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-1-build" podStartSLOduration=3.973246757 podStartE2EDuration="10.291729927s" podCreationTimestamp="2026-03-17 19:00:22 +0000 UTC" firstStartedPulling="2026-03-17 19:00:23.49016614 +0000 UTC m=+767.522777702" lastFinishedPulling="2026-03-17 19:00:29.80864935 +0000 UTC m=+773.841260872" observedRunningTime="2026-03-17 19:00:32.287536355 +0000 UTC m=+776.320147917" watchObservedRunningTime="2026-03-17 19:00:32.291729927 +0000 UTC m=+776.324341489" Mar 17 19:00:32 crc kubenswrapper[5110]: I0317 19:00:32.812135 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.275049 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerName="docker-build" containerID="cri-o://0fb1a78f96260c1c2835b5cb665439730c0368e7d423f678de3943d86c975acf" gracePeriod=30 Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.633280 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.713837 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.713984 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.717506 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-2-global-ca\"" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.717509 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-2-sys-config\"" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.717510 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-2-ca\"" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.816881 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817265 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817303 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817343 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817369 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817539 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817618 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817679 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-push\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817701 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817767 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817807 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lvzp\" (UniqueName: \"kubernetes.io/projected/6e136a9f-ef94-43c2-adb7-36a3208ee96a-kube-api-access-8lvzp\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.817842 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.919735 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.919839 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.919943 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920012 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920183 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920265 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920335 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-push\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920385 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920446 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920457 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920535 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8lvzp\" (UniqueName: \"kubernetes.io/projected/6e136a9f-ef94-43c2-adb7-36a3208ee96a-kube-api-access-8lvzp\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920580 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920704 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.920919 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.921575 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.922232 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.922571 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.923361 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.923358 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.923461 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.923607 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.935360 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-push\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.939953 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:34 crc kubenswrapper[5110]: I0317 19:00:34.940408 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lvzp\" (UniqueName: \"kubernetes.io/projected/6e136a9f-ef94-43c2-adb7-36a3208ee96a-kube-api-access-8lvzp\") pod \"service-telemetry-operator-2-build\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:35 crc kubenswrapper[5110]: I0317 19:00:35.029922 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:00:35 crc kubenswrapper[5110]: I0317 19:00:35.272874 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 17 19:00:35 crc kubenswrapper[5110]: I0317 19:00:35.281168 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerStarted","Data":"0076066cf87733df10ee4b063d807a22209046895dacafca5ad6b8c43f793eee"} Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.294037 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_601ac852-e73a-4c50-a9d6-bbc199c69460/docker-build/0.log" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.294680 5110 generic.go:358] "Generic (PLEG): container finished" podID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerID="0fb1a78f96260c1c2835b5cb665439730c0368e7d423f678de3943d86c975acf" exitCode=1 Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.294771 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"601ac852-e73a-4c50-a9d6-bbc199c69460","Type":"ContainerDied","Data":"0fb1a78f96260c1c2835b5cb665439730c0368e7d423f678de3943d86c975acf"} Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.471863 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_601ac852-e73a-4c50-a9d6-bbc199c69460/docker-build/0.log" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.472274 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.543468 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.547982 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-build-blob-cache\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548078 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-ca-bundles\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548120 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkjrm\" (UniqueName: \"kubernetes.io/projected/601ac852-e73a-4c50-a9d6-bbc199c69460-kube-api-access-vkjrm\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548153 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-system-configs\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548175 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-push\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548209 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-buildcachedir\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548284 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-node-pullsecrets\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548315 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-proxy-ca-bundles\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548385 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-pull\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548411 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-buildworkdir\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548460 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-root\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548479 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-run\") pod \"601ac852-e73a-4c50-a9d6-bbc199c69460\" (UID: \"601ac852-e73a-4c50-a9d6-bbc199c69460\") " Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548542 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548692 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548751 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548875 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.548966 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.549221 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.549303 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.550191 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.550806 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.552307 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.554252 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601ac852-e73a-4c50-a9d6-bbc199c69460-kube-api-access-vkjrm" (OuterVolumeSpecName: "kube-api-access-vkjrm") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "kube-api-access-vkjrm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.554435 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.568785 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "601ac852-e73a-4c50-a9d6-bbc199c69460" (UID: "601ac852-e73a-4c50-a9d6-bbc199c69460"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651248 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651280 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651289 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651298 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vkjrm\" (UniqueName: \"kubernetes.io/projected/601ac852-e73a-4c50-a9d6-bbc199c69460-kube-api-access-vkjrm\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651307 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651316 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651325 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651333 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/601ac852-e73a-4c50-a9d6-bbc199c69460-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651341 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/601ac852-e73a-4c50-a9d6-bbc199c69460-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651349 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/601ac852-e73a-4c50-a9d6-bbc199c69460-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:36 crc kubenswrapper[5110]: I0317 19:00:36.651358 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/601ac852-e73a-4c50-a9d6-bbc199c69460-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.304380 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerStarted","Data":"b18d08f8ae4752e1e348e0fb9c64a50f027174cdbac1bdfe7bf1909f25bd5e15"} Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.307831 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_601ac852-e73a-4c50-a9d6-bbc199c69460/docker-build/0.log" Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.308825 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"601ac852-e73a-4c50-a9d6-bbc199c69460","Type":"ContainerDied","Data":"fafd90ed2609398c9893e648acf918b3913824573295e9157d770cb2e2c145ba"} Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.308906 5110 scope.go:117] "RemoveContainer" containerID="0fb1a78f96260c1c2835b5cb665439730c0368e7d423f678de3943d86c975acf" Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.309142 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.340245 5110 scope.go:117] "RemoveContainer" containerID="2ece55a9c24fb66ddfda9418f3219cc2fdeefa2356e712ca667b672c9ea0e792" Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.362175 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 17 19:00:37 crc kubenswrapper[5110]: I0317 19:00:37.372854 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 17 19:00:38 crc kubenswrapper[5110]: I0317 19:00:38.753361 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" path="/var/lib/kubelet/pods/601ac852-e73a-4c50-a9d6-bbc199c69460/volumes" Mar 17 19:00:42 crc kubenswrapper[5110]: I0317 19:00:42.132694 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:00:42 crc kubenswrapper[5110]: I0317 19:00:42.133321 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:00:45 crc kubenswrapper[5110]: I0317 19:00:45.367578 5110 generic.go:358] "Generic (PLEG): container finished" podID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerID="b18d08f8ae4752e1e348e0fb9c64a50f027174cdbac1bdfe7bf1909f25bd5e15" exitCode=0 Mar 17 19:00:45 crc kubenswrapper[5110]: I0317 19:00:45.367630 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerDied","Data":"b18d08f8ae4752e1e348e0fb9c64a50f027174cdbac1bdfe7bf1909f25bd5e15"} Mar 17 19:00:46 crc kubenswrapper[5110]: I0317 19:00:46.376747 5110 generic.go:358] "Generic (PLEG): container finished" podID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerID="2391ffdad13918e3f0055fa8ca56856a0e92e7a5dd645ddaf7edb38db6aed7e1" exitCode=0 Mar 17 19:00:46 crc kubenswrapper[5110]: I0317 19:00:46.376841 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerDied","Data":"2391ffdad13918e3f0055fa8ca56856a0e92e7a5dd645ddaf7edb38db6aed7e1"} Mar 17 19:00:46 crc kubenswrapper[5110]: I0317 19:00:46.418222 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_6e136a9f-ef94-43c2-adb7-36a3208ee96a/manage-dockerfile/0.log" Mar 17 19:00:47 crc kubenswrapper[5110]: I0317 19:00:47.388891 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerStarted","Data":"5f785491b6ae2ed8cbb1d6f19bd4057f0c2287d491b68fe4fc964e61ad9401fe"} Mar 17 19:00:47 crc kubenswrapper[5110]: I0317 19:00:47.425100 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=13.425075448 podStartE2EDuration="13.425075448s" podCreationTimestamp="2026-03-17 19:00:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:00:47.420415373 +0000 UTC m=+791.453026915" watchObservedRunningTime="2026-03-17 19:00:47.425075448 +0000 UTC m=+791.457686980" Mar 17 19:00:47 crc kubenswrapper[5110]: I0317 19:00:47.912736 5110 scope.go:117] "RemoveContainer" containerID="3d5a1402655aaf02e57e9eecb26b8ba788c0471bd91911a79c43849c55f8d8a5" Mar 17 19:00:48 crc kubenswrapper[5110]: I0317 19:00:48.022366 5110 scope.go:117] "RemoveContainer" containerID="75579bcc7f6b04e5e922a381669b2715f243ef1666a95d2bded26a21b8a6a387" Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.133327 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.134198 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.134280 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.135373 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"399b37230100df387aa59a5f7caddd2a37e880d366524aadcd930dd9995f833d"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.135500 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://399b37230100df387aa59a5f7caddd2a37e880d366524aadcd930dd9995f833d" gracePeriod=600 Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.560911 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="399b37230100df387aa59a5f7caddd2a37e880d366524aadcd930dd9995f833d" exitCode=0 Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.560983 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"399b37230100df387aa59a5f7caddd2a37e880d366524aadcd930dd9995f833d"} Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.561428 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"dd93e17c519f2edf3bf6971330e73cacc5157214f32c427c0e944d936cc0630c"} Mar 17 19:01:12 crc kubenswrapper[5110]: I0317 19:01:12.561454 5110 scope.go:117] "RemoveContainer" containerID="6c003f768dfcdd99f40db6c78311c785574f0d09dbaf1d45833991c2b96d1903" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.148573 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562902-g8ltf"] Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.150894 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerName="manage-dockerfile" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.150927 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerName="manage-dockerfile" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.150978 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerName="docker-build" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.150995 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerName="docker-build" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.151327 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="601ac852-e73a-4c50-a9d6-bbc199c69460" containerName="docker-build" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.175254 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562902-g8ltf"] Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.175389 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.178420 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.178611 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.178505 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.259780 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtrx4\" (UniqueName: \"kubernetes.io/projected/29a01af1-0a8a-4aea-8aad-923a15f691ea-kube-api-access-rtrx4\") pod \"auto-csr-approver-29562902-g8ltf\" (UID: \"29a01af1-0a8a-4aea-8aad-923a15f691ea\") " pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.361526 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rtrx4\" (UniqueName: \"kubernetes.io/projected/29a01af1-0a8a-4aea-8aad-923a15f691ea-kube-api-access-rtrx4\") pod \"auto-csr-approver-29562902-g8ltf\" (UID: \"29a01af1-0a8a-4aea-8aad-923a15f691ea\") " pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.384607 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtrx4\" (UniqueName: \"kubernetes.io/projected/29a01af1-0a8a-4aea-8aad-923a15f691ea-kube-api-access-rtrx4\") pod \"auto-csr-approver-29562902-g8ltf\" (UID: \"29a01af1-0a8a-4aea-8aad-923a15f691ea\") " pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.498181 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.711236 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562902-g8ltf"] Mar 17 19:02:00 crc kubenswrapper[5110]: I0317 19:02:00.885294 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" event={"ID":"29a01af1-0a8a-4aea-8aad-923a15f691ea","Type":"ContainerStarted","Data":"d78015054adc4229ef32f5a7b8d355f0ba5c6e58845d5264c4ed6337504bd020"} Mar 17 19:02:04 crc kubenswrapper[5110]: I0317 19:02:04.199303 5110 generic.go:358] "Generic (PLEG): container finished" podID="29a01af1-0a8a-4aea-8aad-923a15f691ea" containerID="61ccf586dc01b36a60388656b88179e70921e89e1fbc4bbbab0489f02aa702d3" exitCode=0 Mar 17 19:02:04 crc kubenswrapper[5110]: I0317 19:02:04.199410 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" event={"ID":"29a01af1-0a8a-4aea-8aad-923a15f691ea","Type":"ContainerDied","Data":"61ccf586dc01b36a60388656b88179e70921e89e1fbc4bbbab0489f02aa702d3"} Mar 17 19:02:05 crc kubenswrapper[5110]: I0317 19:02:05.420264 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:05 crc kubenswrapper[5110]: I0317 19:02:05.507766 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtrx4\" (UniqueName: \"kubernetes.io/projected/29a01af1-0a8a-4aea-8aad-923a15f691ea-kube-api-access-rtrx4\") pod \"29a01af1-0a8a-4aea-8aad-923a15f691ea\" (UID: \"29a01af1-0a8a-4aea-8aad-923a15f691ea\") " Mar 17 19:02:05 crc kubenswrapper[5110]: I0317 19:02:05.514303 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a01af1-0a8a-4aea-8aad-923a15f691ea-kube-api-access-rtrx4" (OuterVolumeSpecName: "kube-api-access-rtrx4") pod "29a01af1-0a8a-4aea-8aad-923a15f691ea" (UID: "29a01af1-0a8a-4aea-8aad-923a15f691ea"). InnerVolumeSpecName "kube-api-access-rtrx4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:02:05 crc kubenswrapper[5110]: I0317 19:02:05.609041 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rtrx4\" (UniqueName: \"kubernetes.io/projected/29a01af1-0a8a-4aea-8aad-923a15f691ea-kube-api-access-rtrx4\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:06 crc kubenswrapper[5110]: I0317 19:02:06.215963 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" event={"ID":"29a01af1-0a8a-4aea-8aad-923a15f691ea","Type":"ContainerDied","Data":"d78015054adc4229ef32f5a7b8d355f0ba5c6e58845d5264c4ed6337504bd020"} Mar 17 19:02:06 crc kubenswrapper[5110]: I0317 19:02:06.216329 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d78015054adc4229ef32f5a7b8d355f0ba5c6e58845d5264c4ed6337504bd020" Mar 17 19:02:06 crc kubenswrapper[5110]: I0317 19:02:06.216042 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562902-g8ltf" Mar 17 19:02:06 crc kubenswrapper[5110]: I0317 19:02:06.480225 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562896-h6cq4"] Mar 17 19:02:06 crc kubenswrapper[5110]: I0317 19:02:06.484331 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562896-h6cq4"] Mar 17 19:02:06 crc kubenswrapper[5110]: I0317 19:02:06.755376 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8093480b-748b-460c-93cf-ed80e5a3f4ec" path="/var/lib/kubelet/pods/8093480b-748b-460c-93cf-ed80e5a3f4ec/volumes" Mar 17 19:02:16 crc kubenswrapper[5110]: I0317 19:02:16.284080 5110 generic.go:358] "Generic (PLEG): container finished" podID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerID="5f785491b6ae2ed8cbb1d6f19bd4057f0c2287d491b68fe4fc964e61ad9401fe" exitCode=0 Mar 17 19:02:16 crc kubenswrapper[5110]: I0317 19:02:16.284121 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerDied","Data":"5f785491b6ae2ed8cbb1d6f19bd4057f0c2287d491b68fe4fc964e61ad9401fe"} Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.542660 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673466 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-push\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673740 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildcachedir\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673789 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildworkdir\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673806 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-proxy-ca-bundles\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673833 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-blob-cache\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673891 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lvzp\" (UniqueName: \"kubernetes.io/projected/6e136a9f-ef94-43c2-adb7-36a3208ee96a-kube-api-access-8lvzp\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673912 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-ca-bundles\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673932 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-pull\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.673968 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-run\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.674014 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-node-pullsecrets\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.674045 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-root\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.674087 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-system-configs\") pod \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\" (UID: \"6e136a9f-ef94-43c2-adb7-36a3208ee96a\") " Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.674627 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.674817 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.674848 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.680326 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.680457 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.681319 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.681569 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.682842 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e136a9f-ef94-43c2-adb7-36a3208ee96a-kube-api-access-8lvzp" (OuterVolumeSpecName: "kube-api-access-8lvzp") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "kube-api-access-8lvzp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.690593 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.735015 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776165 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776221 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776241 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8lvzp\" (UniqueName: \"kubernetes.io/projected/6e136a9f-ef94-43c2-adb7-36a3208ee96a-kube-api-access-8lvzp\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776258 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776277 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776295 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776311 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776327 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776343 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/6e136a9f-ef94-43c2-adb7-36a3208ee96a-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.776359 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6e136a9f-ef94-43c2-adb7-36a3208ee96a-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.867710 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:17 crc kubenswrapper[5110]: I0317 19:02:17.877990 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:18 crc kubenswrapper[5110]: I0317 19:02:18.303194 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 17 19:02:18 crc kubenswrapper[5110]: I0317 19:02:18.303188 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6e136a9f-ef94-43c2-adb7-36a3208ee96a","Type":"ContainerDied","Data":"0076066cf87733df10ee4b063d807a22209046895dacafca5ad6b8c43f793eee"} Mar 17 19:02:18 crc kubenswrapper[5110]: I0317 19:02:18.303336 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0076066cf87733df10ee4b063d807a22209046895dacafca5ad6b8c43f793eee" Mar 17 19:02:19 crc kubenswrapper[5110]: I0317 19:02:19.598563 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "6e136a9f-ef94-43c2-adb7-36a3208ee96a" (UID: "6e136a9f-ef94-43c2-adb7-36a3208ee96a"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:19 crc kubenswrapper[5110]: I0317 19:02:19.605112 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6e136a9f-ef94-43c2-adb7-36a3208ee96a-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.797157 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798557 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="manage-dockerfile" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798581 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="manage-dockerfile" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798619 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="docker-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798629 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="docker-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798647 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="29a01af1-0a8a-4aea-8aad-923a15f691ea" containerName="oc" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798658 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a01af1-0a8a-4aea-8aad-923a15f691ea" containerName="oc" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798672 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="git-clone" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798681 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="git-clone" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798828 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="29a01af1-0a8a-4aea-8aad-923a15f691ea" containerName="oc" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.798858 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="6e136a9f-ef94-43c2-adb7-36a3208ee96a" containerName="docker-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.914352 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.914485 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.916465 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-1-ca\"" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.916466 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-1-sys-config\"" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.918133 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-1-global-ca\"" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.918232 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949115 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949185 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949217 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdv79\" (UniqueName: \"kubernetes.io/projected/c48f70b3-8b4a-4d54-aa70-6f1463d32619-kube-api-access-hdv79\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949249 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949317 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949391 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949449 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949476 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949506 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949615 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949674 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:22 crc kubenswrapper[5110]: I0317 19:02:22.949736 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-push\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051621 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051740 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051821 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051857 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051891 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051945 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.051979 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052016 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-push\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052152 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052207 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052260 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hdv79\" (UniqueName: \"kubernetes.io/projected/c48f70b3-8b4a-4d54-aa70-6f1463d32619-kube-api-access-hdv79\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052305 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052555 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.052946 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.053433 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.053498 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.053939 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.054006 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.054271 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.054294 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.054735 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.059874 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-push\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.061378 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.073939 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdv79\" (UniqueName: \"kubernetes.io/projected/c48f70b3-8b4a-4d54-aa70-6f1463d32619-kube-api-access-hdv79\") pod \"smart-gateway-operator-1-build\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.232241 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:23 crc kubenswrapper[5110]: I0317 19:02:23.674725 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 17 19:02:24 crc kubenswrapper[5110]: I0317 19:02:24.345969 5110 generic.go:358] "Generic (PLEG): container finished" podID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerID="7d8e116227920b406d4d0a7e68ad5af9e1bfff8085384a16c13f848f0f60572a" exitCode=0 Mar 17 19:02:24 crc kubenswrapper[5110]: I0317 19:02:24.346032 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"c48f70b3-8b4a-4d54-aa70-6f1463d32619","Type":"ContainerDied","Data":"7d8e116227920b406d4d0a7e68ad5af9e1bfff8085384a16c13f848f0f60572a"} Mar 17 19:02:24 crc kubenswrapper[5110]: I0317 19:02:24.346557 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"c48f70b3-8b4a-4d54-aa70-6f1463d32619","Type":"ContainerStarted","Data":"355349c063261a90095d72dbf8616fc576faf5a470a56508fb945b0fb21e5939"} Mar 17 19:02:25 crc kubenswrapper[5110]: I0317 19:02:25.354871 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"c48f70b3-8b4a-4d54-aa70-6f1463d32619","Type":"ContainerStarted","Data":"727b32958d6b15c24f577b881e4ce5073b8960af10dfbda66abf6a3a94fbfad2"} Mar 17 19:02:25 crc kubenswrapper[5110]: I0317 19:02:25.383514 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=3.383491385 podStartE2EDuration="3.383491385s" podCreationTimestamp="2026-03-17 19:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:02:25.381541263 +0000 UTC m=+889.414152785" watchObservedRunningTime="2026-03-17 19:02:25.383491385 +0000 UTC m=+889.416102917" Mar 17 19:02:33 crc kubenswrapper[5110]: I0317 19:02:33.264392 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 17 19:02:33 crc kubenswrapper[5110]: I0317 19:02:33.265573 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerName="docker-build" containerID="cri-o://727b32958d6b15c24f577b881e4ce5073b8960af10dfbda66abf6a3a94fbfad2" gracePeriod=30 Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.419829 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_c48f70b3-8b4a-4d54-aa70-6f1463d32619/docker-build/0.log" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.420919 5110 generic.go:358] "Generic (PLEG): container finished" podID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerID="727b32958d6b15c24f577b881e4ce5073b8960af10dfbda66abf6a3a94fbfad2" exitCode=1 Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.421018 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"c48f70b3-8b4a-4d54-aa70-6f1463d32619","Type":"ContainerDied","Data":"727b32958d6b15c24f577b881e4ce5073b8960af10dfbda66abf6a3a94fbfad2"} Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.862594 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_c48f70b3-8b4a-4d54-aa70-6f1463d32619/docker-build/0.log" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.863169 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.916815 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.917888 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerName="docker-build" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.917915 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerName="docker-build" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.917938 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerName="manage-dockerfile" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.917950 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerName="manage-dockerfile" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.918162 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" containerName="docker-build" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.923296 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.937762 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdv79\" (UniqueName: \"kubernetes.io/projected/c48f70b3-8b4a-4d54-aa70-6f1463d32619-kube-api-access-hdv79\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.937816 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-node-pullsecrets\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.937872 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildcachedir\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.937947 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-system-configs\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.937973 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildworkdir\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938003 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-pull\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938035 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-run\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938095 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-root\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938159 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-proxy-ca-bundles\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938204 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-blob-cache\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938234 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-push\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.938255 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-ca-bundles\") pod \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\" (UID: \"c48f70b3-8b4a-4d54-aa70-6f1463d32619\") " Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.939540 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.939926 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.940420 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.940590 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.940691 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.941347 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.941363 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.941980 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.944699 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-2-ca\"" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.944901 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-2-global-ca\"" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.944997 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-2-sys-config\"" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.948481 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.948504 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.949113 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:02:34 crc kubenswrapper[5110]: I0317 19:02:34.950454 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48f70b3-8b4a-4d54-aa70-6f1463d32619-kube-api-access-hdv79" (OuterVolumeSpecName: "kube-api-access-hdv79") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "kube-api-access-hdv79". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040182 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040246 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040282 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040301 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040333 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040368 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040394 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040466 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cdzd\" (UniqueName: \"kubernetes.io/projected/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-kube-api-access-2cdzd\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040497 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040527 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-push\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040549 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040597 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040653 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040670 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040683 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040696 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040710 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040724 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040737 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040750 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c48f70b3-8b4a-4d54-aa70-6f1463d32619-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040762 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040775 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hdv79\" (UniqueName: \"kubernetes.io/projected/c48f70b3-8b4a-4d54-aa70-6f1463d32619-kube-api-access-hdv79\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.040788 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c48f70b3-8b4a-4d54-aa70-6f1463d32619-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.084700 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "c48f70b3-8b4a-4d54-aa70-6f1463d32619" (UID: "c48f70b3-8b4a-4d54-aa70-6f1463d32619"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.141889 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.141966 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.142704 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.142787 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.142909 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.143670 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.142987 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.143280 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.143692 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.143845 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.143891 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144029 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2cdzd\" (UniqueName: \"kubernetes.io/projected/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-kube-api-access-2cdzd\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144087 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144122 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-push\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144136 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144166 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144173 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144234 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c48f70b3-8b4a-4d54-aa70-6f1463d32619-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144256 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144292 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.144324 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.145804 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.149101 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-push\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.149331 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.162551 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cdzd\" (UniqueName: \"kubernetes.io/projected/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-kube-api-access-2cdzd\") pod \"smart-gateway-operator-2-build\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.292463 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.429765 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_c48f70b3-8b4a-4d54-aa70-6f1463d32619/docker-build/0.log" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.431737 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"c48f70b3-8b4a-4d54-aa70-6f1463d32619","Type":"ContainerDied","Data":"355349c063261a90095d72dbf8616fc576faf5a470a56508fb945b0fb21e5939"} Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.431755 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.431841 5110 scope.go:117] "RemoveContainer" containerID="727b32958d6b15c24f577b881e4ce5073b8960af10dfbda66abf6a3a94fbfad2" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.482608 5110 scope.go:117] "RemoveContainer" containerID="7d8e116227920b406d4d0a7e68ad5af9e1bfff8085384a16c13f848f0f60572a" Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.482745 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.486018 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 17 19:02:35 crc kubenswrapper[5110]: I0317 19:02:35.742366 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 17 19:02:36 crc kubenswrapper[5110]: I0317 19:02:36.440396 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerStarted","Data":"470a084e04348f771ca387b193ee1a43fb216e7b15bffbe92ab42aab565774d0"} Mar 17 19:02:36 crc kubenswrapper[5110]: I0317 19:02:36.440725 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerStarted","Data":"317e8bd327d9dff8672e7f86a60f819a3ea29e6d499e5fe534dcad3f4b7ed1de"} Mar 17 19:02:36 crc kubenswrapper[5110]: I0317 19:02:36.766221 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48f70b3-8b4a-4d54-aa70-6f1463d32619" path="/var/lib/kubelet/pods/c48f70b3-8b4a-4d54-aa70-6f1463d32619/volumes" Mar 17 19:02:37 crc kubenswrapper[5110]: I0317 19:02:37.126768 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:02:37 crc kubenswrapper[5110]: I0317 19:02:37.127358 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:02:37 crc kubenswrapper[5110]: I0317 19:02:37.131697 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:02:37 crc kubenswrapper[5110]: I0317 19:02:37.132146 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:02:37 crc kubenswrapper[5110]: I0317 19:02:37.454796 5110 generic.go:358] "Generic (PLEG): container finished" podID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerID="470a084e04348f771ca387b193ee1a43fb216e7b15bffbe92ab42aab565774d0" exitCode=0 Mar 17 19:02:37 crc kubenswrapper[5110]: I0317 19:02:37.454948 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerDied","Data":"470a084e04348f771ca387b193ee1a43fb216e7b15bffbe92ab42aab565774d0"} Mar 17 19:02:38 crc kubenswrapper[5110]: I0317 19:02:38.472470 5110 generic.go:358] "Generic (PLEG): container finished" podID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerID="ccef025a128e36269d2f9a031e4929c151b3126c5707b0572e4a3799483b2de1" exitCode=0 Mar 17 19:02:38 crc kubenswrapper[5110]: I0317 19:02:38.472588 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerDied","Data":"ccef025a128e36269d2f9a031e4929c151b3126c5707b0572e4a3799483b2de1"} Mar 17 19:02:38 crc kubenswrapper[5110]: I0317 19:02:38.512774 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_9c2f4819-e30b-4497-b3a9-b36dd3022a7e/manage-dockerfile/0.log" Mar 17 19:02:39 crc kubenswrapper[5110]: I0317 19:02:39.483272 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerStarted","Data":"09d62cfe252a5720a02dbeae8491111f292c82951c5385d7004134332128678c"} Mar 17 19:02:39 crc kubenswrapper[5110]: I0317 19:02:39.519214 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=5.519196612 podStartE2EDuration="5.519196612s" podCreationTimestamp="2026-03-17 19:02:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:02:39.516455428 +0000 UTC m=+903.549066960" watchObservedRunningTime="2026-03-17 19:02:39.519196612 +0000 UTC m=+903.551808134" Mar 17 19:02:48 crc kubenswrapper[5110]: I0317 19:02:48.126593 5110 scope.go:117] "RemoveContainer" containerID="bdf416706f8b9b802ac35fcb6bbb69fad04da677b31864ece050af098f442aa4" Mar 17 19:03:12 crc kubenswrapper[5110]: I0317 19:03:12.132786 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:03:12 crc kubenswrapper[5110]: I0317 19:03:12.133491 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:03:42 crc kubenswrapper[5110]: I0317 19:03:42.132551 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:03:42 crc kubenswrapper[5110]: I0317 19:03:42.133142 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:03:46 crc kubenswrapper[5110]: I0317 19:03:46.425166 5110 generic.go:358] "Generic (PLEG): container finished" podID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerID="09d62cfe252a5720a02dbeae8491111f292c82951c5385d7004134332128678c" exitCode=0 Mar 17 19:03:46 crc kubenswrapper[5110]: I0317 19:03:46.425247 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerDied","Data":"09d62cfe252a5720a02dbeae8491111f292c82951c5385d7004134332128678c"} Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.744453 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808313 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-proxy-ca-bundles\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808392 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-push\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808461 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-ca-bundles\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808505 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildcachedir\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808551 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-system-configs\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808645 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808717 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-node-pullsecrets\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808768 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-root\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808801 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-blob-cache\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808842 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildworkdir\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808883 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cdzd\" (UniqueName: \"kubernetes.io/projected/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-kube-api-access-2cdzd\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.808899 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-pull\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.809033 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-run\") pod \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\" (UID: \"9c2f4819-e30b-4497-b3a9-b36dd3022a7e\") " Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.809201 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.809273 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.809337 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.809383 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.809986 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.810000 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.810164 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.810188 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.810208 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.810228 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.816244 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.816304 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.818039 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-kube-api-access-2cdzd" (OuterVolumeSpecName: "kube-api-access-2cdzd") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "kube-api-access-2cdzd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.829967 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.911231 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.911264 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2cdzd\" (UniqueName: \"kubernetes.io/projected/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-kube-api-access-2cdzd\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.911277 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.911298 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:47 crc kubenswrapper[5110]: I0317 19:03:47.911306 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:48 crc kubenswrapper[5110]: I0317 19:03:48.008984 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:03:48 crc kubenswrapper[5110]: I0317 19:03:48.013309 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:48 crc kubenswrapper[5110]: I0317 19:03:48.443250 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 17 19:03:48 crc kubenswrapper[5110]: I0317 19:03:48.443271 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"9c2f4819-e30b-4497-b3a9-b36dd3022a7e","Type":"ContainerDied","Data":"317e8bd327d9dff8672e7f86a60f819a3ea29e6d499e5fe534dcad3f4b7ed1de"} Mar 17 19:03:48 crc kubenswrapper[5110]: I0317 19:03:48.443361 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="317e8bd327d9dff8672e7f86a60f819a3ea29e6d499e5fe534dcad3f4b7ed1de" Mar 17 19:03:49 crc kubenswrapper[5110]: I0317 19:03:49.969914 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9c2f4819-e30b-4497-b3a9-b36dd3022a7e" (UID: "9c2f4819-e30b-4497-b3a9-b36dd3022a7e"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:03:50 crc kubenswrapper[5110]: I0317 19:03:50.038568 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9c2f4819-e30b-4497-b3a9-b36dd3022a7e-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.661047 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.661963 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="manage-dockerfile" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.661976 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="manage-dockerfile" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.661987 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="git-clone" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.661993 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="git-clone" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.662004 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="docker-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.662011 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="docker-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.662104 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="9c2f4819-e30b-4497-b3a9-b36dd3022a7e" containerName="docker-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.705453 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.705577 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.708583 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.708606 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-core-1-global-ca\"" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.708611 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-core-1-ca\"" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.708805 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-core-1-sys-config\"" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.776854 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-run\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.776909 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.776941 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildworkdir\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.776990 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777127 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-system-configs\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777175 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777339 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildcachedir\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777465 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-root\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777551 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-push\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777609 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-pull\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777633 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f4pz\" (UniqueName: \"kubernetes.io/projected/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-kube-api-access-7f4pz\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.777717 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879432 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildcachedir\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879477 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-root\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879502 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-push\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879525 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-pull\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879632 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildcachedir\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879744 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7f4pz\" (UniqueName: \"kubernetes.io/projected/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-kube-api-access-7f4pz\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.879844 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880034 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-run\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880180 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880384 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildworkdir\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880424 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-run\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880436 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880197 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-root\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880562 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880596 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-system-configs\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880613 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.880856 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.881040 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildworkdir\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.881160 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.881403 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.881623 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-system-configs\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.886445 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-pull\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.896853 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-push\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:52 crc kubenswrapper[5110]: I0317 19:03:52.918748 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f4pz\" (UniqueName: \"kubernetes.io/projected/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-kube-api-access-7f4pz\") pod \"sg-core-1-build\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " pod="service-telemetry/sg-core-1-build" Mar 17 19:03:53 crc kubenswrapper[5110]: I0317 19:03:53.024211 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 17 19:03:53 crc kubenswrapper[5110]: I0317 19:03:53.465811 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 17 19:03:53 crc kubenswrapper[5110]: W0317 19:03:53.477005 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod979a9cbb_d41e_41fb_bddd_06f74f88d1bb.slice/crio-8e44623d7a3c21a4a6349c9cacda0146e61e861bdfd8bcedc0e7bf81884e015e WatchSource:0}: Error finding container 8e44623d7a3c21a4a6349c9cacda0146e61e861bdfd8bcedc0e7bf81884e015e: Status 404 returned error can't find the container with id 8e44623d7a3c21a4a6349c9cacda0146e61e861bdfd8bcedc0e7bf81884e015e Mar 17 19:03:53 crc kubenswrapper[5110]: I0317 19:03:53.479766 5110 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 19:03:54 crc kubenswrapper[5110]: I0317 19:03:54.488559 5110 generic.go:358] "Generic (PLEG): container finished" podID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerID="a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3" exitCode=0 Mar 17 19:03:54 crc kubenswrapper[5110]: I0317 19:03:54.488682 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"979a9cbb-d41e-41fb-bddd-06f74f88d1bb","Type":"ContainerDied","Data":"a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3"} Mar 17 19:03:54 crc kubenswrapper[5110]: I0317 19:03:54.488995 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"979a9cbb-d41e-41fb-bddd-06f74f88d1bb","Type":"ContainerStarted","Data":"8e44623d7a3c21a4a6349c9cacda0146e61e861bdfd8bcedc0e7bf81884e015e"} Mar 17 19:03:55 crc kubenswrapper[5110]: I0317 19:03:55.496965 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"979a9cbb-d41e-41fb-bddd-06f74f88d1bb","Type":"ContainerStarted","Data":"ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a"} Mar 17 19:03:55 crc kubenswrapper[5110]: I0317 19:03:55.535682 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=3.535660107 podStartE2EDuration="3.535660107s" podCreationTimestamp="2026-03-17 19:03:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:03:55.528374352 +0000 UTC m=+979.560985904" watchObservedRunningTime="2026-03-17 19:03:55.535660107 +0000 UTC m=+979.568271639" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.152393 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562904-wx68f"] Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.157765 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.160855 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.162948 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562904-wx68f"] Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.166529 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.166682 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.204007 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l72j7\" (UniqueName: \"kubernetes.io/projected/6da3223b-917f-4267-bc6e-e5eb8c595f44-kube-api-access-l72j7\") pod \"auto-csr-approver-29562904-wx68f\" (UID: \"6da3223b-917f-4267-bc6e-e5eb8c595f44\") " pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.305267 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l72j7\" (UniqueName: \"kubernetes.io/projected/6da3223b-917f-4267-bc6e-e5eb8c595f44-kube-api-access-l72j7\") pod \"auto-csr-approver-29562904-wx68f\" (UID: \"6da3223b-917f-4267-bc6e-e5eb8c595f44\") " pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.327690 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l72j7\" (UniqueName: \"kubernetes.io/projected/6da3223b-917f-4267-bc6e-e5eb8c595f44-kube-api-access-l72j7\") pod \"auto-csr-approver-29562904-wx68f\" (UID: \"6da3223b-917f-4267-bc6e-e5eb8c595f44\") " pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.487638 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:00 crc kubenswrapper[5110]: I0317 19:04:00.717005 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562904-wx68f"] Mar 17 19:04:01 crc kubenswrapper[5110]: I0317 19:04:01.537182 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562904-wx68f" event={"ID":"6da3223b-917f-4267-bc6e-e5eb8c595f44","Type":"ContainerStarted","Data":"6cf3cdf52aeca805981029407b7a9f30723550c7ce55618250f73950b023594b"} Mar 17 19:04:02 crc kubenswrapper[5110]: I0317 19:04:02.781176 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 17 19:04:02 crc kubenswrapper[5110]: I0317 19:04:02.781696 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerName="docker-build" containerID="cri-o://ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a" gracePeriod=30 Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.243191 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_979a9cbb-d41e-41fb-bddd-06f74f88d1bb/docker-build/0.log" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.244134 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349168 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-pull\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349227 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-ca-bundles\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349284 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildworkdir\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349323 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-node-pullsecrets\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349354 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-root\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349385 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-run\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349401 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-push\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349431 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-system-configs\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349472 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-proxy-ca-bundles\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349538 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f4pz\" (UniqueName: \"kubernetes.io/projected/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-kube-api-access-7f4pz\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349553 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildcachedir\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.349567 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-blob-cache\") pod \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\" (UID: \"979a9cbb-d41e-41fb-bddd-06f74f88d1bb\") " Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.350792 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.350858 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.351323 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.351787 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.352594 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.352674 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.353562 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.356866 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.359785 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-kube-api-access-7f4pz" (OuterVolumeSpecName: "kube-api-access-7f4pz") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "kube-api-access-7f4pz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.360927 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.439758 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451261 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451296 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451309 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451319 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451327 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451336 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7f4pz\" (UniqueName: \"kubernetes.io/projected/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-kube-api-access-7f4pz\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451344 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451353 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451361 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451369 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.451378 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.465962 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "979a9cbb-d41e-41fb-bddd-06f74f88d1bb" (UID: "979a9cbb-d41e-41fb-bddd-06f74f88d1bb"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.551230 5110 generic.go:358] "Generic (PLEG): container finished" podID="6da3223b-917f-4267-bc6e-e5eb8c595f44" containerID="1c46df711ac7b2cb5092328c8002dce4e55b3585a669047b9eb3fba4bd78da4b" exitCode=0 Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.551369 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562904-wx68f" event={"ID":"6da3223b-917f-4267-bc6e-e5eb8c595f44","Type":"ContainerDied","Data":"1c46df711ac7b2cb5092328c8002dce4e55b3585a669047b9eb3fba4bd78da4b"} Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.551997 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/979a9cbb-d41e-41fb-bddd-06f74f88d1bb-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.557385 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_979a9cbb-d41e-41fb-bddd-06f74f88d1bb/docker-build/0.log" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.558413 5110 generic.go:358] "Generic (PLEG): container finished" podID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerID="ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a" exitCode=1 Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.558453 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"979a9cbb-d41e-41fb-bddd-06f74f88d1bb","Type":"ContainerDied","Data":"ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a"} Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.558490 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.558521 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"979a9cbb-d41e-41fb-bddd-06f74f88d1bb","Type":"ContainerDied","Data":"8e44623d7a3c21a4a6349c9cacda0146e61e861bdfd8bcedc0e7bf81884e015e"} Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.558546 5110 scope.go:117] "RemoveContainer" containerID="ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.598236 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.602967 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.609377 5110 scope.go:117] "RemoveContainer" containerID="a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.661606 5110 scope.go:117] "RemoveContainer" containerID="ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a" Mar 17 19:04:03 crc kubenswrapper[5110]: E0317 19:04:03.662028 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a\": container with ID starting with ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a not found: ID does not exist" containerID="ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.662135 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a"} err="failed to get container status \"ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a\": rpc error: code = NotFound desc = could not find container \"ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a\": container with ID starting with ddac7a930f791679f739c76cdc1a5011b015079525c9e41bdc0c403a0a0fba0a not found: ID does not exist" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.662172 5110 scope.go:117] "RemoveContainer" containerID="a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3" Mar 17 19:04:03 crc kubenswrapper[5110]: E0317 19:04:03.662491 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3\": container with ID starting with a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3 not found: ID does not exist" containerID="a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3" Mar 17 19:04:03 crc kubenswrapper[5110]: I0317 19:04:03.662520 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3"} err="failed to get container status \"a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3\": rpc error: code = NotFound desc = could not find container \"a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3\": container with ID starting with a1382d584df1a621925d69ac565134a35949254878d55bd7d31e7f90b93a4fc3 not found: ID does not exist" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.392590 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.393228 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerName="docker-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.393242 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerName="docker-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.393270 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerName="manage-dockerfile" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.393276 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerName="manage-dockerfile" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.393374 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" containerName="docker-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.438609 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.438721 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.440827 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-core-2-global-ca\"" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.441305 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-core-2-ca\"" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.441719 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-core-2-sys-config\"" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.442832 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463229 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-push\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463271 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-pull\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463305 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-run\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463338 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxgb7\" (UniqueName: \"kubernetes.io/projected/403dfa7a-7eb3-4209-822a-658a9982411a-kube-api-access-qxgb7\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463356 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-system-configs\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463377 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463394 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463410 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-buildcachedir\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463427 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463463 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-buildworkdir\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463481 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.463511 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-root\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.564200 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565098 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565245 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565321 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-buildcachedir\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565378 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565402 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565423 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-buildcachedir\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565546 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-buildworkdir\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565617 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.565723 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-root\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566209 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-buildworkdir\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566240 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-push\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566278 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-root\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566312 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566344 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-pull\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566693 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-run\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566810 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qxgb7\" (UniqueName: \"kubernetes.io/projected/403dfa7a-7eb3-4209-822a-658a9982411a-kube-api-access-qxgb7\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.566952 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-system-configs\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.567583 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-run\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.567719 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.567768 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-system-configs\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.572536 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-push\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.572864 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-pull\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.593561 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxgb7\" (UniqueName: \"kubernetes.io/projected/403dfa7a-7eb3-4209-822a-658a9982411a-kube-api-access-qxgb7\") pod \"sg-core-2-build\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.753827 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="979a9cbb-d41e-41fb-bddd-06f74f88d1bb" path="/var/lib/kubelet/pods/979a9cbb-d41e-41fb-bddd-06f74f88d1bb/volumes" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.764636 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.818705 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.871259 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l72j7\" (UniqueName: \"kubernetes.io/projected/6da3223b-917f-4267-bc6e-e5eb8c595f44-kube-api-access-l72j7\") pod \"6da3223b-917f-4267-bc6e-e5eb8c595f44\" (UID: \"6da3223b-917f-4267-bc6e-e5eb8c595f44\") " Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.877266 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da3223b-917f-4267-bc6e-e5eb8c595f44-kube-api-access-l72j7" (OuterVolumeSpecName: "kube-api-access-l72j7") pod "6da3223b-917f-4267-bc6e-e5eb8c595f44" (UID: "6da3223b-917f-4267-bc6e-e5eb8c595f44"). InnerVolumeSpecName "kube-api-access-l72j7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:04:04 crc kubenswrapper[5110]: I0317 19:04:04.973437 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l72j7\" (UniqueName: \"kubernetes.io/projected/6da3223b-917f-4267-bc6e-e5eb8c595f44-kube-api-access-l72j7\") on node \"crc\" DevicePath \"\"" Mar 17 19:04:05 crc kubenswrapper[5110]: W0317 19:04:05.164642 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod403dfa7a_7eb3_4209_822a_658a9982411a.slice/crio-9406a5a255ceeb5ecc59c47e3fe201b5f8a2aa6fb11f2701075f63dad25582bc WatchSource:0}: Error finding container 9406a5a255ceeb5ecc59c47e3fe201b5f8a2aa6fb11f2701075f63dad25582bc: Status 404 returned error can't find the container with id 9406a5a255ceeb5ecc59c47e3fe201b5f8a2aa6fb11f2701075f63dad25582bc Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.172732 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.578781 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562904-wx68f" Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.578844 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562904-wx68f" event={"ID":"6da3223b-917f-4267-bc6e-e5eb8c595f44","Type":"ContainerDied","Data":"6cf3cdf52aeca805981029407b7a9f30723550c7ce55618250f73950b023594b"} Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.578926 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cf3cdf52aeca805981029407b7a9f30723550c7ce55618250f73950b023594b" Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.582155 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerStarted","Data":"53156e56cf34f60b91e19b5bb3628eb68dc2ebd3a3db8d076153680c19fa42ba"} Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.582213 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerStarted","Data":"9406a5a255ceeb5ecc59c47e3fe201b5f8a2aa6fb11f2701075f63dad25582bc"} Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.875395 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562898-5lc6h"] Mar 17 19:04:05 crc kubenswrapper[5110]: I0317 19:04:05.881188 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562898-5lc6h"] Mar 17 19:04:06 crc kubenswrapper[5110]: I0317 19:04:06.592228 5110 generic.go:358] "Generic (PLEG): container finished" podID="403dfa7a-7eb3-4209-822a-658a9982411a" containerID="53156e56cf34f60b91e19b5bb3628eb68dc2ebd3a3db8d076153680c19fa42ba" exitCode=0 Mar 17 19:04:06 crc kubenswrapper[5110]: I0317 19:04:06.592283 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerDied","Data":"53156e56cf34f60b91e19b5bb3628eb68dc2ebd3a3db8d076153680c19fa42ba"} Mar 17 19:04:06 crc kubenswrapper[5110]: I0317 19:04:06.757049 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e77e6a5-f972-41ad-9d00-d86d43c0176d" path="/var/lib/kubelet/pods/3e77e6a5-f972-41ad-9d00-d86d43c0176d/volumes" Mar 17 19:04:07 crc kubenswrapper[5110]: I0317 19:04:07.601776 5110 generic.go:358] "Generic (PLEG): container finished" podID="403dfa7a-7eb3-4209-822a-658a9982411a" containerID="a2e197ee7ae30d549a7253b253dc121f6a9d107b2a511d51e76e7d92982c9cc7" exitCode=0 Mar 17 19:04:07 crc kubenswrapper[5110]: I0317 19:04:07.601865 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerDied","Data":"a2e197ee7ae30d549a7253b253dc121f6a9d107b2a511d51e76e7d92982c9cc7"} Mar 17 19:04:07 crc kubenswrapper[5110]: I0317 19:04:07.638482 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_403dfa7a-7eb3-4209-822a-658a9982411a/manage-dockerfile/0.log" Mar 17 19:04:08 crc kubenswrapper[5110]: I0317 19:04:08.612714 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerStarted","Data":"8d759c33d8f3daf097e05c730286a55922b7c186aed32c6c49e141abf19dc5c2"} Mar 17 19:04:08 crc kubenswrapper[5110]: I0317 19:04:08.644847 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=4.644829969 podStartE2EDuration="4.644829969s" podCreationTimestamp="2026-03-17 19:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:04:08.643911995 +0000 UTC m=+992.676523517" watchObservedRunningTime="2026-03-17 19:04:08.644829969 +0000 UTC m=+992.677441491" Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.132514 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.132823 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.132873 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.133489 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd93e17c519f2edf3bf6971330e73cacc5157214f32c427c0e944d936cc0630c"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.133585 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://dd93e17c519f2edf3bf6971330e73cacc5157214f32c427c0e944d936cc0630c" gracePeriod=600 Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.664860 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="dd93e17c519f2edf3bf6971330e73cacc5157214f32c427c0e944d936cc0630c" exitCode=0 Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.665014 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"dd93e17c519f2edf3bf6971330e73cacc5157214f32c427c0e944d936cc0630c"} Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.665457 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"0938f11ace42511813dad951668253bd6932f3e91eca29ddfdba80f92fcd4637"} Mar 17 19:04:12 crc kubenswrapper[5110]: I0317 19:04:12.665504 5110 scope.go:117] "RemoveContainer" containerID="399b37230100df387aa59a5f7caddd2a37e880d366524aadcd930dd9995f833d" Mar 17 19:04:48 crc kubenswrapper[5110]: I0317 19:04:48.296038 5110 scope.go:117] "RemoveContainer" containerID="c28b96f67244aad2ab7aeeeafb8aa8e4547040532f58078ded1ff354716699fa" Mar 17 19:05:26 crc kubenswrapper[5110]: I0317 19:05:26.444497 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qz95x"] Mar 17 19:05:26 crc kubenswrapper[5110]: I0317 19:05:26.447469 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6da3223b-917f-4267-bc6e-e5eb8c595f44" containerName="oc" Mar 17 19:05:26 crc kubenswrapper[5110]: I0317 19:05:26.447589 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da3223b-917f-4267-bc6e-e5eb8c595f44" containerName="oc" Mar 17 19:05:26 crc kubenswrapper[5110]: I0317 19:05:26.448130 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="6da3223b-917f-4267-bc6e-e5eb8c595f44" containerName="oc" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.597133 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz95x"] Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.597389 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.684446 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c6e43d-9340-4b90-879e-9508669131f1-catalog-content\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.684562 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psl7d\" (UniqueName: \"kubernetes.io/projected/d9c6e43d-9340-4b90-879e-9508669131f1-kube-api-access-psl7d\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.684921 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c6e43d-9340-4b90-879e-9508669131f1-utilities\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.786001 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c6e43d-9340-4b90-879e-9508669131f1-utilities\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.786328 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c6e43d-9340-4b90-879e-9508669131f1-catalog-content\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.786507 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-psl7d\" (UniqueName: \"kubernetes.io/projected/d9c6e43d-9340-4b90-879e-9508669131f1-kube-api-access-psl7d\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.786801 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c6e43d-9340-4b90-879e-9508669131f1-utilities\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.787080 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c6e43d-9340-4b90-879e-9508669131f1-catalog-content\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.811030 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-psl7d\" (UniqueName: \"kubernetes.io/projected/d9c6e43d-9340-4b90-879e-9508669131f1-kube-api-access-psl7d\") pod \"certified-operators-qz95x\" (UID: \"d9c6e43d-9340-4b90-879e-9508669131f1\") " pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:27 crc kubenswrapper[5110]: I0317 19:05:27.920636 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.120786 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz95x"] Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.173175 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz95x" event={"ID":"d9c6e43d-9340-4b90-879e-9508669131f1","Type":"ContainerStarted","Data":"42107cb9797cb29e0bcee670a2bd1c3f9f5e945f01214269b3a374add475d216"} Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.638328 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rdc9p"] Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.692351 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdc9p"] Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.692631 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.798230 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-utilities\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.798327 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-catalog-content\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.798453 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f58p2\" (UniqueName: \"kubernetes.io/projected/65d1d877-ecce-4145-8b15-d6a0f3a60fda-kube-api-access-f58p2\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.899935 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-catalog-content\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.900281 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f58p2\" (UniqueName: \"kubernetes.io/projected/65d1d877-ecce-4145-8b15-d6a0f3a60fda-kube-api-access-f58p2\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.900358 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-utilities\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.900597 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-catalog-content\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.900785 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-utilities\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:28 crc kubenswrapper[5110]: I0317 19:05:28.931454 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f58p2\" (UniqueName: \"kubernetes.io/projected/65d1d877-ecce-4145-8b15-d6a0f3a60fda-kube-api-access-f58p2\") pod \"redhat-operators-rdc9p\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:29 crc kubenswrapper[5110]: I0317 19:05:29.010580 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:29 crc kubenswrapper[5110]: I0317 19:05:29.180599 5110 generic.go:358] "Generic (PLEG): container finished" podID="d9c6e43d-9340-4b90-879e-9508669131f1" containerID="f6b641537f0d689d8b3b00d8bbf78940ad8e800bc59536abdd8503fd515e70a7" exitCode=0 Mar 17 19:05:29 crc kubenswrapper[5110]: I0317 19:05:29.180778 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz95x" event={"ID":"d9c6e43d-9340-4b90-879e-9508669131f1","Type":"ContainerDied","Data":"f6b641537f0d689d8b3b00d8bbf78940ad8e800bc59536abdd8503fd515e70a7"} Mar 17 19:05:29 crc kubenswrapper[5110]: I0317 19:05:29.424132 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdc9p"] Mar 17 19:05:30 crc kubenswrapper[5110]: I0317 19:05:30.187453 5110 generic.go:358] "Generic (PLEG): container finished" podID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerID="8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405" exitCode=0 Mar 17 19:05:30 crc kubenswrapper[5110]: I0317 19:05:30.187498 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdc9p" event={"ID":"65d1d877-ecce-4145-8b15-d6a0f3a60fda","Type":"ContainerDied","Data":"8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405"} Mar 17 19:05:30 crc kubenswrapper[5110]: I0317 19:05:30.187849 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdc9p" event={"ID":"65d1d877-ecce-4145-8b15-d6a0f3a60fda","Type":"ContainerStarted","Data":"ab166d21d3e040f8be4f8972b597495d72abb74e4e107dedc3a68848afa7ee28"} Mar 17 19:05:33 crc kubenswrapper[5110]: I0317 19:05:33.208322 5110 generic.go:358] "Generic (PLEG): container finished" podID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerID="ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed" exitCode=0 Mar 17 19:05:33 crc kubenswrapper[5110]: I0317 19:05:33.208907 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdc9p" event={"ID":"65d1d877-ecce-4145-8b15-d6a0f3a60fda","Type":"ContainerDied","Data":"ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed"} Mar 17 19:05:35 crc kubenswrapper[5110]: I0317 19:05:35.221434 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdc9p" event={"ID":"65d1d877-ecce-4145-8b15-d6a0f3a60fda","Type":"ContainerStarted","Data":"14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3"} Mar 17 19:05:35 crc kubenswrapper[5110]: I0317 19:05:35.223643 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz95x" event={"ID":"d9c6e43d-9340-4b90-879e-9508669131f1","Type":"ContainerStarted","Data":"75c829b42ad6984a2dfd3728f5bec30ed7e8049083e6591f53de26d85499c2c0"} Mar 17 19:05:35 crc kubenswrapper[5110]: I0317 19:05:35.240147 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rdc9p" podStartSLOduration=5.762977109 podStartE2EDuration="7.240129773s" podCreationTimestamp="2026-03-17 19:05:28 +0000 UTC" firstStartedPulling="2026-03-17 19:05:30.188421174 +0000 UTC m=+1074.221032696" lastFinishedPulling="2026-03-17 19:05:31.665573838 +0000 UTC m=+1075.698185360" observedRunningTime="2026-03-17 19:05:35.239508947 +0000 UTC m=+1079.272120479" watchObservedRunningTime="2026-03-17 19:05:35.240129773 +0000 UTC m=+1079.272741305" Mar 17 19:05:36 crc kubenswrapper[5110]: I0317 19:05:36.230943 5110 generic.go:358] "Generic (PLEG): container finished" podID="d9c6e43d-9340-4b90-879e-9508669131f1" containerID="75c829b42ad6984a2dfd3728f5bec30ed7e8049083e6591f53de26d85499c2c0" exitCode=0 Mar 17 19:05:36 crc kubenswrapper[5110]: I0317 19:05:36.230993 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz95x" event={"ID":"d9c6e43d-9340-4b90-879e-9508669131f1","Type":"ContainerDied","Data":"75c829b42ad6984a2dfd3728f5bec30ed7e8049083e6591f53de26d85499c2c0"} Mar 17 19:05:37 crc kubenswrapper[5110]: I0317 19:05:37.244977 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz95x" event={"ID":"d9c6e43d-9340-4b90-879e-9508669131f1","Type":"ContainerStarted","Data":"23d7e1d186ec661f9e9c5a17057c81907f3b0b4281f0a9a3e4a3dd20b3b24a2b"} Mar 17 19:05:37 crc kubenswrapper[5110]: I0317 19:05:37.266650 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qz95x" podStartSLOduration=5.554111782 podStartE2EDuration="11.266628926s" podCreationTimestamp="2026-03-17 19:05:26 +0000 UTC" firstStartedPulling="2026-03-17 19:05:29.181556178 +0000 UTC m=+1073.214167710" lastFinishedPulling="2026-03-17 19:05:34.894073332 +0000 UTC m=+1078.926684854" observedRunningTime="2026-03-17 19:05:37.260189644 +0000 UTC m=+1081.292801176" watchObservedRunningTime="2026-03-17 19:05:37.266628926 +0000 UTC m=+1081.299240448" Mar 17 19:05:37 crc kubenswrapper[5110]: I0317 19:05:37.921494 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:37 crc kubenswrapper[5110]: I0317 19:05:37.921876 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:37 crc kubenswrapper[5110]: I0317 19:05:37.968015 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:39 crc kubenswrapper[5110]: I0317 19:05:39.011281 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:39 crc kubenswrapper[5110]: I0317 19:05:39.011347 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:40 crc kubenswrapper[5110]: I0317 19:05:40.067473 5110 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rdc9p" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="registry-server" probeResult="failure" output=< Mar 17 19:05:40 crc kubenswrapper[5110]: timeout: failed to connect service ":50051" within 1s Mar 17 19:05:40 crc kubenswrapper[5110]: > Mar 17 19:05:49 crc kubenswrapper[5110]: I0317 19:05:49.077115 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:49 crc kubenswrapper[5110]: I0317 19:05:49.125310 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:49 crc kubenswrapper[5110]: I0317 19:05:49.297351 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qz95x" Mar 17 19:05:49 crc kubenswrapper[5110]: I0317 19:05:49.319108 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdc9p"] Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.343233 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rdc9p" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="registry-server" containerID="cri-o://14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3" gracePeriod=2 Mar 17 19:05:50 crc kubenswrapper[5110]: E0317 19:05:50.467990 5110 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d1d877_ecce_4145_8b15_d6a0f3a60fda.slice/crio-14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3.scope\": RecentStats: unable to find data in memory cache]" Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.724754 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.822587 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f58p2\" (UniqueName: \"kubernetes.io/projected/65d1d877-ecce-4145-8b15-d6a0f3a60fda-kube-api-access-f58p2\") pod \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.822690 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-catalog-content\") pod \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.822818 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-utilities\") pod \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\" (UID: \"65d1d877-ecce-4145-8b15-d6a0f3a60fda\") " Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.823877 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-utilities" (OuterVolumeSpecName: "utilities") pod "65d1d877-ecce-4145-8b15-d6a0f3a60fda" (UID: "65d1d877-ecce-4145-8b15-d6a0f3a60fda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.844603 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d1d877-ecce-4145-8b15-d6a0f3a60fda-kube-api-access-f58p2" (OuterVolumeSpecName: "kube-api-access-f58p2") pod "65d1d877-ecce-4145-8b15-d6a0f3a60fda" (UID: "65d1d877-ecce-4145-8b15-d6a0f3a60fda"). InnerVolumeSpecName "kube-api-access-f58p2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.925373 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.925421 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-f58p2\" (UniqueName: \"kubernetes.io/projected/65d1d877-ecce-4145-8b15-d6a0f3a60fda-kube-api-access-f58p2\") on node \"crc\" DevicePath \"\"" Mar 17 19:05:50 crc kubenswrapper[5110]: I0317 19:05:50.939590 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65d1d877-ecce-4145-8b15-d6a0f3a60fda" (UID: "65d1d877-ecce-4145-8b15-d6a0f3a60fda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.026320 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d1d877-ecce-4145-8b15-d6a0f3a60fda-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.345085 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz95x"] Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.353190 5110 generic.go:358] "Generic (PLEG): container finished" podID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerID="14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3" exitCode=0 Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.353275 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdc9p" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.353349 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdc9p" event={"ID":"65d1d877-ecce-4145-8b15-d6a0f3a60fda","Type":"ContainerDied","Data":"14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3"} Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.353427 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdc9p" event={"ID":"65d1d877-ecce-4145-8b15-d6a0f3a60fda","Type":"ContainerDied","Data":"ab166d21d3e040f8be4f8972b597495d72abb74e4e107dedc3a68848afa7ee28"} Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.353459 5110 scope.go:117] "RemoveContainer" containerID="14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.382111 5110 scope.go:117] "RemoveContainer" containerID="ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.401203 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdc9p"] Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.402393 5110 scope.go:117] "RemoveContainer" containerID="8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.406434 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rdc9p"] Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.417131 5110 scope.go:117] "RemoveContainer" containerID="14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3" Mar 17 19:05:51 crc kubenswrapper[5110]: E0317 19:05:51.417668 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3\": container with ID starting with 14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3 not found: ID does not exist" containerID="14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.417703 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3"} err="failed to get container status \"14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3\": rpc error: code = NotFound desc = could not find container \"14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3\": container with ID starting with 14a6c3713af0f456d0672985a83d1c3ab1f38cfe60cef9840818dfb2b2c4dba3 not found: ID does not exist" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.417732 5110 scope.go:117] "RemoveContainer" containerID="ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed" Mar 17 19:05:51 crc kubenswrapper[5110]: E0317 19:05:51.418038 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed\": container with ID starting with ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed not found: ID does not exist" containerID="ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.418095 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed"} err="failed to get container status \"ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed\": rpc error: code = NotFound desc = could not find container \"ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed\": container with ID starting with ae31dc0d52ed0373e565eb14d4728d9270dc8a09b7b845449d49c908c395b2ed not found: ID does not exist" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.418125 5110 scope.go:117] "RemoveContainer" containerID="8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405" Mar 17 19:05:51 crc kubenswrapper[5110]: E0317 19:05:51.418582 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405\": container with ID starting with 8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405 not found: ID does not exist" containerID="8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.418619 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405"} err="failed to get container status \"8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405\": rpc error: code = NotFound desc = could not find container \"8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405\": container with ID starting with 8e2ab5c25c63ebb1e35525dc7ace551719984abbcfee433596a4758d6ef17405 not found: ID does not exist" Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.714395 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pfbcf"] Mar 17 19:05:51 crc kubenswrapper[5110]: I0317 19:05:51.714680 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pfbcf" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="registry-server" containerID="cri-o://42e7b9b2ea6d754a5623e9e547bcd92a370638378f4c1bffef9e57065f4f924e" gracePeriod=2 Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.366570 5110 generic.go:358] "Generic (PLEG): container finished" podID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerID="42e7b9b2ea6d754a5623e9e547bcd92a370638378f4c1bffef9e57065f4f924e" exitCode=0 Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.366632 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerDied","Data":"42e7b9b2ea6d754a5623e9e547bcd92a370638378f4c1bffef9e57065f4f924e"} Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.620449 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.751046 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-utilities\") pod \"0861862c-403b-4786-bc1d-0c9d9fab15b5\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.751190 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mj26\" (UniqueName: \"kubernetes.io/projected/0861862c-403b-4786-bc1d-0c9d9fab15b5-kube-api-access-9mj26\") pod \"0861862c-403b-4786-bc1d-0c9d9fab15b5\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.751352 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-catalog-content\") pod \"0861862c-403b-4786-bc1d-0c9d9fab15b5\" (UID: \"0861862c-403b-4786-bc1d-0c9d9fab15b5\") " Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.754840 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-utilities" (OuterVolumeSpecName: "utilities") pod "0861862c-403b-4786-bc1d-0c9d9fab15b5" (UID: "0861862c-403b-4786-bc1d-0c9d9fab15b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.760739 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0861862c-403b-4786-bc1d-0c9d9fab15b5-kube-api-access-9mj26" (OuterVolumeSpecName: "kube-api-access-9mj26") pod "0861862c-403b-4786-bc1d-0c9d9fab15b5" (UID: "0861862c-403b-4786-bc1d-0c9d9fab15b5"). InnerVolumeSpecName "kube-api-access-9mj26". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.766265 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" path="/var/lib/kubelet/pods/65d1d877-ecce-4145-8b15-d6a0f3a60fda/volumes" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.821394 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0861862c-403b-4786-bc1d-0c9d9fab15b5" (UID: "0861862c-403b-4786-bc1d-0c9d9fab15b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.854402 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.854446 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0861862c-403b-4786-bc1d-0c9d9fab15b5-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 19:05:52 crc kubenswrapper[5110]: I0317 19:05:52.854461 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9mj26\" (UniqueName: \"kubernetes.io/projected/0861862c-403b-4786-bc1d-0c9d9fab15b5-kube-api-access-9mj26\") on node \"crc\" DevicePath \"\"" Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.385293 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfbcf" event={"ID":"0861862c-403b-4786-bc1d-0c9d9fab15b5","Type":"ContainerDied","Data":"c8e3054df075c38686996efbc315dccdb925300191e33f692944d70821661f13"} Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.385485 5110 scope.go:117] "RemoveContainer" containerID="42e7b9b2ea6d754a5623e9e547bcd92a370638378f4c1bffef9e57065f4f924e" Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.385508 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfbcf" Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.419286 5110 scope.go:117] "RemoveContainer" containerID="5624f47da017bf23fc43ad49f2c445de17871aa9cfc2ad2266c690145d0cd46f" Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.424531 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pfbcf"] Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.432195 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pfbcf"] Mar 17 19:05:53 crc kubenswrapper[5110]: I0317 19:05:53.438825 5110 scope.go:117] "RemoveContainer" containerID="b4f7d9e5953553869617712e07b22a9b9b8c49e19ccf09d006620c94336b54d1" Mar 17 19:05:54 crc kubenswrapper[5110]: I0317 19:05:54.751297 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" path="/var/lib/kubelet/pods/0861862c-403b-4786-bc1d-0c9d9fab15b5/volumes" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.132440 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562906-vm2mw"] Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.132984 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="extract-content" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.132996 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="extract-content" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133019 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="extract-utilities" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133025 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="extract-utilities" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133033 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="extract-content" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133040 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="extract-content" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133068 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="registry-server" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133073 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="registry-server" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133092 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="registry-server" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133097 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="registry-server" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133111 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="extract-utilities" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133116 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="extract-utilities" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133208 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="0861862c-403b-4786-bc1d-0c9d9fab15b5" containerName="registry-server" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.133220 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="65d1d877-ecce-4145-8b15-d6a0f3a60fda" containerName="registry-server" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.224509 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562906-vm2mw"] Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.224656 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.226617 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.227481 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.230371 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.270724 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swjlp\" (UniqueName: \"kubernetes.io/projected/7c0e3708-d07f-4103-b22b-ba3fa96d0aed-kube-api-access-swjlp\") pod \"auto-csr-approver-29562906-vm2mw\" (UID: \"7c0e3708-d07f-4103-b22b-ba3fa96d0aed\") " pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.371874 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-swjlp\" (UniqueName: \"kubernetes.io/projected/7c0e3708-d07f-4103-b22b-ba3fa96d0aed-kube-api-access-swjlp\") pod \"auto-csr-approver-29562906-vm2mw\" (UID: \"7c0e3708-d07f-4103-b22b-ba3fa96d0aed\") " pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.403780 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-swjlp\" (UniqueName: \"kubernetes.io/projected/7c0e3708-d07f-4103-b22b-ba3fa96d0aed-kube-api-access-swjlp\") pod \"auto-csr-approver-29562906-vm2mw\" (UID: \"7c0e3708-d07f-4103-b22b-ba3fa96d0aed\") " pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:00 crc kubenswrapper[5110]: I0317 19:06:00.547726 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:01 crc kubenswrapper[5110]: I0317 19:06:01.027261 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562906-vm2mw"] Mar 17 19:06:01 crc kubenswrapper[5110]: I0317 19:06:01.439579 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" event={"ID":"7c0e3708-d07f-4103-b22b-ba3fa96d0aed","Type":"ContainerStarted","Data":"aada5040e41889cdf02f7f2f90e1f5aeff1d1eb93ec43860d48f97b0986a3098"} Mar 17 19:06:03 crc kubenswrapper[5110]: I0317 19:06:03.455393 5110 generic.go:358] "Generic (PLEG): container finished" podID="7c0e3708-d07f-4103-b22b-ba3fa96d0aed" containerID="35d758205c4e3e118fd55027d0804dff72a46218879b69c30d269fbc907d5524" exitCode=0 Mar 17 19:06:03 crc kubenswrapper[5110]: I0317 19:06:03.455643 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" event={"ID":"7c0e3708-d07f-4103-b22b-ba3fa96d0aed","Type":"ContainerDied","Data":"35d758205c4e3e118fd55027d0804dff72a46218879b69c30d269fbc907d5524"} Mar 17 19:06:04 crc kubenswrapper[5110]: I0317 19:06:04.694427 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:04 crc kubenswrapper[5110]: I0317 19:06:04.833452 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swjlp\" (UniqueName: \"kubernetes.io/projected/7c0e3708-d07f-4103-b22b-ba3fa96d0aed-kube-api-access-swjlp\") pod \"7c0e3708-d07f-4103-b22b-ba3fa96d0aed\" (UID: \"7c0e3708-d07f-4103-b22b-ba3fa96d0aed\") " Mar 17 19:06:04 crc kubenswrapper[5110]: I0317 19:06:04.840436 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0e3708-d07f-4103-b22b-ba3fa96d0aed-kube-api-access-swjlp" (OuterVolumeSpecName: "kube-api-access-swjlp") pod "7c0e3708-d07f-4103-b22b-ba3fa96d0aed" (UID: "7c0e3708-d07f-4103-b22b-ba3fa96d0aed"). InnerVolumeSpecName "kube-api-access-swjlp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:06:04 crc kubenswrapper[5110]: I0317 19:06:04.935562 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-swjlp\" (UniqueName: \"kubernetes.io/projected/7c0e3708-d07f-4103-b22b-ba3fa96d0aed-kube-api-access-swjlp\") on node \"crc\" DevicePath \"\"" Mar 17 19:06:05 crc kubenswrapper[5110]: I0317 19:06:05.473832 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" event={"ID":"7c0e3708-d07f-4103-b22b-ba3fa96d0aed","Type":"ContainerDied","Data":"aada5040e41889cdf02f7f2f90e1f5aeff1d1eb93ec43860d48f97b0986a3098"} Mar 17 19:06:05 crc kubenswrapper[5110]: I0317 19:06:05.473927 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aada5040e41889cdf02f7f2f90e1f5aeff1d1eb93ec43860d48f97b0986a3098" Mar 17 19:06:05 crc kubenswrapper[5110]: I0317 19:06:05.474033 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562906-vm2mw" Mar 17 19:06:05 crc kubenswrapper[5110]: I0317 19:06:05.747221 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562900-hsptz"] Mar 17 19:06:05 crc kubenswrapper[5110]: I0317 19:06:05.751551 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562900-hsptz"] Mar 17 19:06:06 crc kubenswrapper[5110]: I0317 19:06:06.754378 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ca21359-d4f9-4720-8e0f-21c5b449d62f" path="/var/lib/kubelet/pods/7ca21359-d4f9-4720-8e0f-21c5b449d62f/volumes" Mar 17 19:06:12 crc kubenswrapper[5110]: I0317 19:06:12.132986 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:06:12 crc kubenswrapper[5110]: I0317 19:06:12.133368 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:06:42 crc kubenswrapper[5110]: I0317 19:06:42.133169 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:06:42 crc kubenswrapper[5110]: I0317 19:06:42.133853 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:06:48 crc kubenswrapper[5110]: I0317 19:06:48.441581 5110 scope.go:117] "RemoveContainer" containerID="67391f7570881da2e65abbeed016414d864b17053df32bf0a3cf7a5fcfde1932" Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.133305 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.133985 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.134038 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.134686 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0938f11ace42511813dad951668253bd6932f3e91eca29ddfdba80f92fcd4637"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.134745 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://0938f11ace42511813dad951668253bd6932f3e91eca29ddfdba80f92fcd4637" gracePeriod=600 Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.959445 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="0938f11ace42511813dad951668253bd6932f3e91eca29ddfdba80f92fcd4637" exitCode=0 Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.959534 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"0938f11ace42511813dad951668253bd6932f3e91eca29ddfdba80f92fcd4637"} Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.960078 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"cb0b6120b2825b80ee94ecacb872639139788855e7496255ef0eb30ec1f502af"} Mar 17 19:07:12 crc kubenswrapper[5110]: I0317 19:07:12.960096 5110 scope.go:117] "RemoveContainer" containerID="dd93e17c519f2edf3bf6971330e73cacc5157214f32c427c0e944d936cc0630c" Mar 17 19:07:37 crc kubenswrapper[5110]: I0317 19:07:37.401136 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:07:37 crc kubenswrapper[5110]: I0317 19:07:37.402967 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:07:37 crc kubenswrapper[5110]: I0317 19:07:37.408662 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:07:37 crc kubenswrapper[5110]: I0317 19:07:37.409003 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:07:38 crc kubenswrapper[5110]: I0317 19:07:38.126183 5110 generic.go:358] "Generic (PLEG): container finished" podID="403dfa7a-7eb3-4209-822a-658a9982411a" containerID="8d759c33d8f3daf097e05c730286a55922b7c186aed32c6c49e141abf19dc5c2" exitCode=0 Mar 17 19:07:38 crc kubenswrapper[5110]: I0317 19:07:38.126255 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerDied","Data":"8d759c33d8f3daf097e05c730286a55922b7c186aed32c6c49e141abf19dc5c2"} Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.367002 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.418871 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-push\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.418931 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-proxy-ca-bundles\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419039 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-ca-bundles\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419108 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-root\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419125 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-build-blob-cache\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419153 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-pull\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419171 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-buildworkdir\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419197 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-buildcachedir\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419226 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-system-configs\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419268 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxgb7\" (UniqueName: \"kubernetes.io/projected/403dfa7a-7eb3-4209-822a-658a9982411a-kube-api-access-qxgb7\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419296 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-run\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419328 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-node-pullsecrets\") pod \"403dfa7a-7eb3-4209-822a-658a9982411a\" (UID: \"403dfa7a-7eb3-4209-822a-658a9982411a\") " Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419581 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.419589 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.420102 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.420216 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.420623 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.421047 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.436378 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/403dfa7a-7eb3-4209-822a-658a9982411a-kube-api-access-qxgb7" (OuterVolumeSpecName: "kube-api-access-qxgb7") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "kube-api-access-qxgb7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.436393 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.438122 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.446226 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520601 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520654 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520683 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520692 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520700 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520708 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qxgb7\" (UniqueName: \"kubernetes.io/projected/403dfa7a-7eb3-4209-822a-658a9982411a-kube-api-access-qxgb7\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520716 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520724 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/403dfa7a-7eb3-4209-822a-658a9982411a-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520732 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/403dfa7a-7eb3-4209-822a-658a9982411a-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.520741 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/403dfa7a-7eb3-4209-822a-658a9982411a-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.758240 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:39 crc kubenswrapper[5110]: I0317 19:07:39.826554 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:40 crc kubenswrapper[5110]: I0317 19:07:40.141620 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"403dfa7a-7eb3-4209-822a-658a9982411a","Type":"ContainerDied","Data":"9406a5a255ceeb5ecc59c47e3fe201b5f8a2aa6fb11f2701075f63dad25582bc"} Mar 17 19:07:40 crc kubenswrapper[5110]: I0317 19:07:40.141665 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9406a5a255ceeb5ecc59c47e3fe201b5f8a2aa6fb11f2701075f63dad25582bc" Mar 17 19:07:40 crc kubenswrapper[5110]: I0317 19:07:40.141744 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 17 19:07:42 crc kubenswrapper[5110]: I0317 19:07:42.368347 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "403dfa7a-7eb3-4209-822a-658a9982411a" (UID: "403dfa7a-7eb3-4209-822a-658a9982411a"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:42 crc kubenswrapper[5110]: I0317 19:07:42.467742 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/403dfa7a-7eb3-4209-822a-658a9982411a-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.992891 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994478 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7c0e3708-d07f-4103-b22b-ba3fa96d0aed" containerName="oc" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994512 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e3708-d07f-4103-b22b-ba3fa96d0aed" containerName="oc" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994534 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="git-clone" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994548 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="git-clone" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994582 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="docker-build" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994595 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="docker-build" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994623 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="manage-dockerfile" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994637 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="manage-dockerfile" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994914 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="7c0e3708-d07f-4103-b22b-ba3fa96d0aed" containerName="oc" Mar 17 19:07:43 crc kubenswrapper[5110]: I0317 19:07:43.994947 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="403dfa7a-7eb3-4209-822a-658a9982411a" containerName="docker-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.000448 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.002362 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.003922 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-bridge-1-global-ca\"" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.004266 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-bridge-1-sys-config\"" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.004694 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-bridge-1-ca\"" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.008495 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091279 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091324 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091374 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091471 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-push\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091537 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-pull\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091583 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091624 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091655 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091694 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091728 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091813 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5lwh\" (UniqueName: \"kubernetes.io/projected/8403e329-3a1f-40dc-835a-a72c4b16a573-kube-api-access-b5lwh\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.091894 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193656 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193735 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193780 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193833 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-push\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193879 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-pull\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193919 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.193962 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194003 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194046 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194166 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194328 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b5lwh\" (UniqueName: \"kubernetes.io/projected/8403e329-3a1f-40dc-835a-a72c4b16a573-kube-api-access-b5lwh\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194456 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194861 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.194466 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.195240 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.195448 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.195806 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.195942 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.196790 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.197277 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.197735 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.201998 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-pull\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.202533 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-push\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.218187 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5lwh\" (UniqueName: \"kubernetes.io/projected/8403e329-3a1f-40dc-835a-a72c4b16a573-kube-api-access-b5lwh\") pod \"sg-bridge-1-build\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.370313 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:44 crc kubenswrapper[5110]: I0317 19:07:44.672106 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 17 19:07:45 crc kubenswrapper[5110]: I0317 19:07:45.178258 5110 generic.go:358] "Generic (PLEG): container finished" podID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerID="de58c28d2ea8e429ae66acd28161736dd3adf73435ba9faedbd16b165d7d13b6" exitCode=0 Mar 17 19:07:45 crc kubenswrapper[5110]: I0317 19:07:45.178618 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"8403e329-3a1f-40dc-835a-a72c4b16a573","Type":"ContainerDied","Data":"de58c28d2ea8e429ae66acd28161736dd3adf73435ba9faedbd16b165d7d13b6"} Mar 17 19:07:45 crc kubenswrapper[5110]: I0317 19:07:45.178941 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"8403e329-3a1f-40dc-835a-a72c4b16a573","Type":"ContainerStarted","Data":"3fb181173ca46c6dfa48047e8f3f695eda9df1909a9ddd18cf3144a808a45194"} Mar 17 19:07:46 crc kubenswrapper[5110]: I0317 19:07:46.191113 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"8403e329-3a1f-40dc-835a-a72c4b16a573","Type":"ContainerStarted","Data":"350134e3836062644ab4eeadfd6c035e7d877aed77f65775c736f0c901040f8f"} Mar 17 19:07:46 crc kubenswrapper[5110]: I0317 19:07:46.221545 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=3.22150772 podStartE2EDuration="3.22150772s" podCreationTimestamp="2026-03-17 19:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:07:46.218423417 +0000 UTC m=+1210.251034979" watchObservedRunningTime="2026-03-17 19:07:46.22150772 +0000 UTC m=+1210.254119242" Mar 17 19:07:54 crc kubenswrapper[5110]: I0317 19:07:54.226635 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 17 19:07:54 crc kubenswrapper[5110]: I0317 19:07:54.227503 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/sg-bridge-1-build" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerName="docker-build" containerID="cri-o://350134e3836062644ab4eeadfd6c035e7d877aed77f65775c736f0c901040f8f" gracePeriod=30 Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.261263 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_8403e329-3a1f-40dc-835a-a72c4b16a573/docker-build/0.log" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.261786 5110 generic.go:358] "Generic (PLEG): container finished" podID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerID="350134e3836062644ab4eeadfd6c035e7d877aed77f65775c736f0c901040f8f" exitCode=1 Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.261899 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"8403e329-3a1f-40dc-835a-a72c4b16a573","Type":"ContainerDied","Data":"350134e3836062644ab4eeadfd6c035e7d877aed77f65775c736f0c901040f8f"} Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.753690 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_8403e329-3a1f-40dc-835a-a72c4b16a573/docker-build/0.log" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.754205 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.850829 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.851515 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerName="manage-dockerfile" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.851538 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerName="manage-dockerfile" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.851558 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerName="docker-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.851564 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerName="docker-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.851711 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" containerName="docker-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.858218 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.861296 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-bridge-2-global-ca\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.861296 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-bridge-2-sys-config\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.861659 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"sg-bridge-2-ca\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.867237 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-buildcachedir\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.867536 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.867763 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5lwh\" (UniqueName: \"kubernetes.io/projected/8403e329-3a1f-40dc-835a-a72c4b16a573-kube-api-access-b5lwh\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.869114 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-proxy-ca-bundles\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.869352 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-pull\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.869507 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-ca-bundles\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.869767 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.870265 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.870610 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-buildworkdir\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.871395 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.872253 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.872497 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-build-blob-cache\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874129 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-system-configs\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874161 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-run\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874187 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-node-pullsecrets\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874210 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-push\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874241 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-root\") pod \"8403e329-3a1f-40dc-835a-a72c4b16a573\" (UID: \"8403e329-3a1f-40dc-835a-a72c4b16a573\") " Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874387 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874417 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874448 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874504 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874550 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874621 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874679 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874702 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874732 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874754 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-push\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874788 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-pull\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.874755 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.875424 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.875549 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877346 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877429 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87gkw\" (UniqueName: \"kubernetes.io/projected/7425a770-32e7-4b71-b2d4-7fd233519a7a-kube-api-access-87gkw\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877492 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877507 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877520 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877531 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877542 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8403e329-3a1f-40dc-835a-a72c4b16a573-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877553 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877565 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.877575 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8403e329-3a1f-40dc-835a-a72c4b16a573-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.879822 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.883483 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8403e329-3a1f-40dc-835a-a72c4b16a573-kube-api-access-b5lwh" (OuterVolumeSpecName: "kube-api-access-b5lwh") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "kube-api-access-b5lwh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.931943 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.978826 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-pull\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979091 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979206 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-87gkw\" (UniqueName: \"kubernetes.io/projected/7425a770-32e7-4b71-b2d4-7fd233519a7a-kube-api-access-87gkw\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979300 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979402 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979491 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979571 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979657 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979747 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979822 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979895 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.979960 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-push\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.980076 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-b5lwh\" (UniqueName: \"kubernetes.io/projected/8403e329-3a1f-40dc-835a-a72c4b16a573-kube-api-access-b5lwh\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.980153 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.980217 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/8403e329-3a1f-40dc-835a-a72c4b16a573-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.981336 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.981551 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.981581 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.981948 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.982246 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.982252 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.982344 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.983012 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.983898 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.985326 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-pull\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.985300 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-push\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:55 crc kubenswrapper[5110]: I0317 19:07:55.996960 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-87gkw\" (UniqueName: \"kubernetes.io/projected/7425a770-32e7-4b71-b2d4-7fd233519a7a-kube-api-access-87gkw\") pod \"sg-bridge-2-build\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.174882 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "8403e329-3a1f-40dc-835a-a72c4b16a573" (UID: "8403e329-3a1f-40dc-835a-a72c4b16a573"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.177265 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.183124 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8403e329-3a1f-40dc-835a-a72c4b16a573-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.276921 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_8403e329-3a1f-40dc-835a-a72c4b16a573/docker-build/0.log" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.277986 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"8403e329-3a1f-40dc-835a-a72c4b16a573","Type":"ContainerDied","Data":"3fb181173ca46c6dfa48047e8f3f695eda9df1909a9ddd18cf3144a808a45194"} Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.278040 5110 scope.go:117] "RemoveContainer" containerID="350134e3836062644ab4eeadfd6c035e7d877aed77f65775c736f0c901040f8f" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.278229 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.328285 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.333380 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.336360 5110 scope.go:117] "RemoveContainer" containerID="de58c28d2ea8e429ae66acd28161736dd3adf73435ba9faedbd16b165d7d13b6" Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.418915 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 17 19:07:56 crc kubenswrapper[5110]: I0317 19:07:56.755746 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8403e329-3a1f-40dc-835a-a72c4b16a573" path="/var/lib/kubelet/pods/8403e329-3a1f-40dc-835a-a72c4b16a573/volumes" Mar 17 19:07:57 crc kubenswrapper[5110]: I0317 19:07:57.288041 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerStarted","Data":"80cfe024bcb0864d3f71bf9bd993f8e9a706f0e1452d596d1c25e92a1daa19d3"} Mar 17 19:07:57 crc kubenswrapper[5110]: I0317 19:07:57.288495 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerStarted","Data":"86f42bacecf20e246934eba17c574049a4e9598ef4521bcf4744d1d8c4b5727d"} Mar 17 19:07:58 crc kubenswrapper[5110]: I0317 19:07:58.301029 5110 generic.go:358] "Generic (PLEG): container finished" podID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerID="80cfe024bcb0864d3f71bf9bd993f8e9a706f0e1452d596d1c25e92a1daa19d3" exitCode=0 Mar 17 19:07:58 crc kubenswrapper[5110]: I0317 19:07:58.301379 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerDied","Data":"80cfe024bcb0864d3f71bf9bd993f8e9a706f0e1452d596d1c25e92a1daa19d3"} Mar 17 19:07:59 crc kubenswrapper[5110]: I0317 19:07:59.313750 5110 generic.go:358] "Generic (PLEG): container finished" podID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerID="64be0741c44ef305aa9cd591c1422538b9dbaa6f9a99b966e51a8d21db35c29a" exitCode=0 Mar 17 19:07:59 crc kubenswrapper[5110]: I0317 19:07:59.313826 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerDied","Data":"64be0741c44ef305aa9cd591c1422538b9dbaa6f9a99b966e51a8d21db35c29a"} Mar 17 19:07:59 crc kubenswrapper[5110]: I0317 19:07:59.348613 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_7425a770-32e7-4b71-b2d4-7fd233519a7a/manage-dockerfile/0.log" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.145731 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562908-xzlpd"] Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.151620 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.154837 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.156254 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.156407 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562908-xzlpd"] Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.156653 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.250245 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sb6p\" (UniqueName: \"kubernetes.io/projected/11779d6a-7707-4288-aaa3-8ae9411fb3e7-kube-api-access-6sb6p\") pod \"auto-csr-approver-29562908-xzlpd\" (UID: \"11779d6a-7707-4288-aaa3-8ae9411fb3e7\") " pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.323827 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerStarted","Data":"b83af30e218d0da6f0c49e265a06e68c8bd6efeedbdc6ef6eaddcb795b07baa9"} Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.351340 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6sb6p\" (UniqueName: \"kubernetes.io/projected/11779d6a-7707-4288-aaa3-8ae9411fb3e7-kube-api-access-6sb6p\") pod \"auto-csr-approver-29562908-xzlpd\" (UID: \"11779d6a-7707-4288-aaa3-8ae9411fb3e7\") " pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.351862 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.351839673 podStartE2EDuration="5.351839673s" podCreationTimestamp="2026-03-17 19:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:08:00.347350902 +0000 UTC m=+1224.379962424" watchObservedRunningTime="2026-03-17 19:08:00.351839673 +0000 UTC m=+1224.384451215" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.377333 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sb6p\" (UniqueName: \"kubernetes.io/projected/11779d6a-7707-4288-aaa3-8ae9411fb3e7-kube-api-access-6sb6p\") pod \"auto-csr-approver-29562908-xzlpd\" (UID: \"11779d6a-7707-4288-aaa3-8ae9411fb3e7\") " pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.477940 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:00 crc kubenswrapper[5110]: I0317 19:08:00.890079 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562908-xzlpd"] Mar 17 19:08:01 crc kubenswrapper[5110]: I0317 19:08:01.334398 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" event={"ID":"11779d6a-7707-4288-aaa3-8ae9411fb3e7","Type":"ContainerStarted","Data":"c876269e9a7fb6a301ede1fa8d0f9947c3cc342d1498397058b690457a2824a7"} Mar 17 19:08:02 crc kubenswrapper[5110]: I0317 19:08:02.343838 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" event={"ID":"11779d6a-7707-4288-aaa3-8ae9411fb3e7","Type":"ContainerStarted","Data":"007780b833e3d22bcc72a3bfa388d586fd85b933df1b8afe805cea9fc0ddf97b"} Mar 17 19:08:03 crc kubenswrapper[5110]: I0317 19:08:03.352287 5110 generic.go:358] "Generic (PLEG): container finished" podID="11779d6a-7707-4288-aaa3-8ae9411fb3e7" containerID="007780b833e3d22bcc72a3bfa388d586fd85b933df1b8afe805cea9fc0ddf97b" exitCode=0 Mar 17 19:08:03 crc kubenswrapper[5110]: I0317 19:08:03.353196 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" event={"ID":"11779d6a-7707-4288-aaa3-8ae9411fb3e7","Type":"ContainerDied","Data":"007780b833e3d22bcc72a3bfa388d586fd85b933df1b8afe805cea9fc0ddf97b"} Mar 17 19:08:04 crc kubenswrapper[5110]: I0317 19:08:04.599189 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:04 crc kubenswrapper[5110]: I0317 19:08:04.664330 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sb6p\" (UniqueName: \"kubernetes.io/projected/11779d6a-7707-4288-aaa3-8ae9411fb3e7-kube-api-access-6sb6p\") pod \"11779d6a-7707-4288-aaa3-8ae9411fb3e7\" (UID: \"11779d6a-7707-4288-aaa3-8ae9411fb3e7\") " Mar 17 19:08:04 crc kubenswrapper[5110]: I0317 19:08:04.674310 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11779d6a-7707-4288-aaa3-8ae9411fb3e7-kube-api-access-6sb6p" (OuterVolumeSpecName: "kube-api-access-6sb6p") pod "11779d6a-7707-4288-aaa3-8ae9411fb3e7" (UID: "11779d6a-7707-4288-aaa3-8ae9411fb3e7"). InnerVolumeSpecName "kube-api-access-6sb6p". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:08:04 crc kubenswrapper[5110]: I0317 19:08:04.766302 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6sb6p\" (UniqueName: \"kubernetes.io/projected/11779d6a-7707-4288-aaa3-8ae9411fb3e7-kube-api-access-6sb6p\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:04 crc kubenswrapper[5110]: I0317 19:08:04.993107 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562902-g8ltf"] Mar 17 19:08:04 crc kubenswrapper[5110]: I0317 19:08:04.997184 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562902-g8ltf"] Mar 17 19:08:05 crc kubenswrapper[5110]: I0317 19:08:05.369351 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" event={"ID":"11779d6a-7707-4288-aaa3-8ae9411fb3e7","Type":"ContainerDied","Data":"c876269e9a7fb6a301ede1fa8d0f9947c3cc342d1498397058b690457a2824a7"} Mar 17 19:08:05 crc kubenswrapper[5110]: I0317 19:08:05.369420 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c876269e9a7fb6a301ede1fa8d0f9947c3cc342d1498397058b690457a2824a7" Mar 17 19:08:05 crc kubenswrapper[5110]: I0317 19:08:05.370000 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562908-xzlpd" Mar 17 19:08:06 crc kubenswrapper[5110]: I0317 19:08:06.753596 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a01af1-0a8a-4aea-8aad-923a15f691ea" path="/var/lib/kubelet/pods/29a01af1-0a8a-4aea-8aad-923a15f691ea/volumes" Mar 17 19:08:47 crc kubenswrapper[5110]: I0317 19:08:47.716828 5110 generic.go:358] "Generic (PLEG): container finished" podID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerID="b83af30e218d0da6f0c49e265a06e68c8bd6efeedbdc6ef6eaddcb795b07baa9" exitCode=0 Mar 17 19:08:47 crc kubenswrapper[5110]: I0317 19:08:47.716906 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerDied","Data":"b83af30e218d0da6f0c49e265a06e68c8bd6efeedbdc6ef6eaddcb795b07baa9"} Mar 17 19:08:48 crc kubenswrapper[5110]: I0317 19:08:48.589177 5110 scope.go:117] "RemoveContainer" containerID="61ccf586dc01b36a60388656b88179e70921e89e1fbc4bbbab0489f02aa702d3" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.014967 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.047854 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-root\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.047940 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-push\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.047989 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-blob-cache\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048018 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildcachedir\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048118 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildworkdir\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048185 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-proxy-ca-bundles\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048222 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-node-pullsecrets\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048243 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-system-configs\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048281 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-run\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048302 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-pull\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048375 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-ca-bundles\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.048400 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87gkw\" (UniqueName: \"kubernetes.io/projected/7425a770-32e7-4b71-b2d4-7fd233519a7a-kube-api-access-87gkw\") pod \"7425a770-32e7-4b71-b2d4-7fd233519a7a\" (UID: \"7425a770-32e7-4b71-b2d4-7fd233519a7a\") " Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.050614 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.051573 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.052117 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.052355 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.052556 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.053557 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.055348 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.056513 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.061566 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7425a770-32e7-4b71-b2d4-7fd233519a7a-kube-api-access-87gkw" (OuterVolumeSpecName: "kube-api-access-87gkw") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "kube-api-access-87gkw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.067343 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149517 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149548 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149557 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149565 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149573 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149605 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149614 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-87gkw\" (UniqueName: \"kubernetes.io/projected/7425a770-32e7-4b71-b2d4-7fd233519a7a-kube-api-access-87gkw\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149623 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/7425a770-32e7-4b71-b2d4-7fd233519a7a-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149632 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.149640 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.181394 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.251029 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.732835 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"7425a770-32e7-4b71-b2d4-7fd233519a7a","Type":"ContainerDied","Data":"86f42bacecf20e246934eba17c574049a4e9598ef4521bcf4744d1d8c4b5727d"} Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.733181 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86f42bacecf20e246934eba17c574049a4e9598ef4521bcf4744d1d8c4b5727d" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.733129 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.874271 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "7425a770-32e7-4b71-b2d4-7fd233519a7a" (UID: "7425a770-32e7-4b71-b2d4-7fd233519a7a"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:08:49 crc kubenswrapper[5110]: I0317 19:08:49.968482 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7425a770-32e7-4b71-b2d4-7fd233519a7a-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.449230 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450665 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="docker-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450711 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="docker-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450759 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="manage-dockerfile" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450772 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="manage-dockerfile" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450808 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="git-clone" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450824 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="git-clone" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450841 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="11779d6a-7707-4288-aaa3-8ae9411fb3e7" containerName="oc" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.450857 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="11779d6a-7707-4288-aaa3-8ae9411fb3e7" containerName="oc" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.451146 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="7425a770-32e7-4b71-b2d4-7fd233519a7a" containerName="docker-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.451202 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="11779d6a-7707-4288-aaa3-8ae9411fb3e7" containerName="oc" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.662612 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.662734 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.666088 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-webhook-snmp-1-global-ca\"" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.666204 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-webhook-snmp-1-sys-config\"" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.666440 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.667393 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-webhook-snmp-1-ca\"" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722648 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722717 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722765 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722801 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722831 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722870 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722897 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722922 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722957 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.722982 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.723010 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.723041 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vdsd\" (UniqueName: \"kubernetes.io/projected/9ffe5f63-f130-4871-a548-68ba5a3f59af-kube-api-access-6vdsd\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.824535 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.824649 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.824744 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.824861 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.824869 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.824923 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.825003 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.825176 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.825220 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.825291 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.825961 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.826024 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.825365 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.826150 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.826245 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.826742 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.827083 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.827167 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.827270 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6vdsd\" (UniqueName: \"kubernetes.io/projected/9ffe5f63-f130-4871-a548-68ba5a3f59af-kube-api-access-6vdsd\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.827595 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.828425 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.834133 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.834675 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.856858 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vdsd\" (UniqueName: \"kubernetes.io/projected/9ffe5f63-f130-4871-a548-68ba5a3f59af-kube-api-access-6vdsd\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:53 crc kubenswrapper[5110]: I0317 19:08:53.989766 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:08:54 crc kubenswrapper[5110]: I0317 19:08:54.273650 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 17 19:08:54 crc kubenswrapper[5110]: I0317 19:08:54.278902 5110 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 19:08:54 crc kubenswrapper[5110]: I0317 19:08:54.792703 5110 generic.go:358] "Generic (PLEG): container finished" podID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerID="fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3" exitCode=0 Mar 17 19:08:54 crc kubenswrapper[5110]: I0317 19:08:54.792833 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"9ffe5f63-f130-4871-a548-68ba5a3f59af","Type":"ContainerDied","Data":"fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3"} Mar 17 19:08:54 crc kubenswrapper[5110]: I0317 19:08:54.793323 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"9ffe5f63-f130-4871-a548-68ba5a3f59af","Type":"ContainerStarted","Data":"8d2b27011403dc9b5c3a47ae881202f0c310baba1b1c308efbebeeae8f55b09f"} Mar 17 19:08:55 crc kubenswrapper[5110]: I0317 19:08:55.807049 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"9ffe5f63-f130-4871-a548-68ba5a3f59af","Type":"ContainerStarted","Data":"ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9"} Mar 17 19:08:55 crc kubenswrapper[5110]: I0317 19:08:55.839899 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=2.839872775 podStartE2EDuration="2.839872775s" podCreationTimestamp="2026-03-17 19:08:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:08:55.839003362 +0000 UTC m=+1279.871614924" watchObservedRunningTime="2026-03-17 19:08:55.839872775 +0000 UTC m=+1279.872484337" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.051349 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.052591 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerName="docker-build" containerID="cri-o://ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9" gracePeriod=30 Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.442283 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_9ffe5f63-f130-4871-a548-68ba5a3f59af/docker-build/0.log" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.443015 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.495451 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-run\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.495716 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildworkdir\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.495897 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-proxy-ca-bundles\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496041 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-pull\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496185 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-root\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496262 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-blob-cache\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496399 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-node-pullsecrets\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496503 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vdsd\" (UniqueName: \"kubernetes.io/projected/9ffe5f63-f130-4871-a548-68ba5a3f59af-kube-api-access-6vdsd\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496586 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildcachedir\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496521 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.496611 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.497118 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.497232 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.497335 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.497401 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.498264 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.501752 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.502134 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ffe5f63-f130-4871-a548-68ba5a3f59af-kube-api-access-6vdsd" (OuterVolumeSpecName: "kube-api-access-6vdsd") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "kube-api-access-6vdsd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.554868 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598368 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-push\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598427 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-ca-bundles\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598454 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-system-configs\") pod \"9ffe5f63-f130-4871-a548-68ba5a3f59af\" (UID: \"9ffe5f63-f130-4871-a548-68ba5a3f59af\") " Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598718 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598740 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598752 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598763 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598776 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.598793 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6vdsd\" (UniqueName: \"kubernetes.io/projected/9ffe5f63-f130-4871-a548-68ba5a3f59af-kube-api-access-6vdsd\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.599097 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.599456 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.602046 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.700465 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/9ffe5f63-f130-4871-a548-68ba5a3f59af-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.700503 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.700514 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9ffe5f63-f130-4871-a548-68ba5a3f59af-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.879923 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_9ffe5f63-f130-4871-a548-68ba5a3f59af/docker-build/0.log" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.880570 5110 generic.go:358] "Generic (PLEG): container finished" podID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerID="ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9" exitCode=1 Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.880734 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"9ffe5f63-f130-4871-a548-68ba5a3f59af","Type":"ContainerDied","Data":"ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9"} Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.880772 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"9ffe5f63-f130-4871-a548-68ba5a3f59af","Type":"ContainerDied","Data":"8d2b27011403dc9b5c3a47ae881202f0c310baba1b1c308efbebeeae8f55b09f"} Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.880794 5110 scope.go:117] "RemoveContainer" containerID="ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.880947 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.898490 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9ffe5f63-f130-4871-a548-68ba5a3f59af" (UID: "9ffe5f63-f130-4871-a548-68ba5a3f59af"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.904414 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9ffe5f63-f130-4871-a548-68ba5a3f59af-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.907142 5110 scope.go:117] "RemoveContainer" containerID="fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.975049 5110 scope.go:117] "RemoveContainer" containerID="ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9" Mar 17 19:09:04 crc kubenswrapper[5110]: E0317 19:09:04.975411 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9\": container with ID starting with ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9 not found: ID does not exist" containerID="ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.975458 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9"} err="failed to get container status \"ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9\": rpc error: code = NotFound desc = could not find container \"ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9\": container with ID starting with ba48c59f03270b9b01eee5b4f52dc61cdd831cd782a09d684eb9d4a875937bb9 not found: ID does not exist" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.975482 5110 scope.go:117] "RemoveContainer" containerID="fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3" Mar 17 19:09:04 crc kubenswrapper[5110]: E0317 19:09:04.975837 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3\": container with ID starting with fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3 not found: ID does not exist" containerID="fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3" Mar 17 19:09:04 crc kubenswrapper[5110]: I0317 19:09:04.975942 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3"} err="failed to get container status \"fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3\": rpc error: code = NotFound desc = could not find container \"fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3\": container with ID starting with fe6e4f1f565165f8b8ec48350916ddaed9f66a2dd1114d6d2c3d9d046d4f5df3 not found: ID does not exist" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.224513 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.229359 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.651359 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.657646 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerName="docker-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.657745 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerName="docker-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.657831 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerName="manage-dockerfile" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.657907 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerName="manage-dockerfile" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.658124 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" containerName="docker-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.769687 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.769858 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.772360 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-webhook-snmp-2-sys-config\"" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.772523 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-webhook-snmp-2-global-ca\"" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.772542 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-webhook-snmp-2-ca\"" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.772765 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"builder-dockercfg-mpnmp\"" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.816620 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.816737 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.816764 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.816827 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.816847 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.816941 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.817032 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.817160 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.817202 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.817218 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxtpn\" (UniqueName: \"kubernetes.io/projected/c9bfda01-07e3-4539-a8ed-857e07b64db3-kube-api-access-wxtpn\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.817365 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.817527 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.918975 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919025 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919085 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919122 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919259 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919310 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919335 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919371 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919411 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919442 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919451 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919471 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919496 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wxtpn\" (UniqueName: \"kubernetes.io/projected/c9bfda01-07e3-4539-a8ed-857e07b64db3-kube-api-access-wxtpn\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.919853 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.920221 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.920352 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.920548 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.920656 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.920881 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.920955 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.921126 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.924492 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.927620 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:05 crc kubenswrapper[5110]: I0317 19:09:05.936195 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxtpn\" (UniqueName: \"kubernetes.io/projected/c9bfda01-07e3-4539-a8ed-857e07b64db3-kube-api-access-wxtpn\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:06 crc kubenswrapper[5110]: I0317 19:09:06.081378 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:09:06 crc kubenswrapper[5110]: I0317 19:09:06.287076 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 17 19:09:06 crc kubenswrapper[5110]: I0317 19:09:06.756817 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ffe5f63-f130-4871-a548-68ba5a3f59af" path="/var/lib/kubelet/pods/9ffe5f63-f130-4871-a548-68ba5a3f59af/volumes" Mar 17 19:09:06 crc kubenswrapper[5110]: I0317 19:09:06.899481 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerStarted","Data":"2c59a3a75e30acd174e4c2e615f4ca01b5d8d8304872149d40df1257f2242865"} Mar 17 19:09:06 crc kubenswrapper[5110]: I0317 19:09:06.899527 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerStarted","Data":"3ffd32a68a1d873d28c7f4ec8ca7ccc7d4948ceef8f050b732b4be6fddb9eba6"} Mar 17 19:09:07 crc kubenswrapper[5110]: I0317 19:09:07.909309 5110 generic.go:358] "Generic (PLEG): container finished" podID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerID="2c59a3a75e30acd174e4c2e615f4ca01b5d8d8304872149d40df1257f2242865" exitCode=0 Mar 17 19:09:07 crc kubenswrapper[5110]: I0317 19:09:07.909370 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerDied","Data":"2c59a3a75e30acd174e4c2e615f4ca01b5d8d8304872149d40df1257f2242865"} Mar 17 19:09:08 crc kubenswrapper[5110]: I0317 19:09:08.920995 5110 generic.go:358] "Generic (PLEG): container finished" podID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerID="5618db10770abee1a1f55ccf9c2351b31218744b6a2b168597374731bfd054b5" exitCode=0 Mar 17 19:09:08 crc kubenswrapper[5110]: I0317 19:09:08.921107 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerDied","Data":"5618db10770abee1a1f55ccf9c2351b31218744b6a2b168597374731bfd054b5"} Mar 17 19:09:08 crc kubenswrapper[5110]: I0317 19:09:08.974214 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_c9bfda01-07e3-4539-a8ed-857e07b64db3/manage-dockerfile/0.log" Mar 17 19:09:09 crc kubenswrapper[5110]: I0317 19:09:09.935383 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerStarted","Data":"5ff1bacb27c99674a6a50ec6cc1f83c5e2a2e3a0c1eaab88d9226377971cea6f"} Mar 17 19:09:09 crc kubenswrapper[5110]: I0317 19:09:09.969255 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=4.969233147 podStartE2EDuration="4.969233147s" podCreationTimestamp="2026-03-17 19:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:09:09.962244297 +0000 UTC m=+1293.994855839" watchObservedRunningTime="2026-03-17 19:09:09.969233147 +0000 UTC m=+1294.001844689" Mar 17 19:09:12 crc kubenswrapper[5110]: I0317 19:09:12.133444 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:09:12 crc kubenswrapper[5110]: I0317 19:09:12.133562 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:09:42 crc kubenswrapper[5110]: I0317 19:09:42.132774 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:09:42 crc kubenswrapper[5110]: I0317 19:09:42.133623 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.136852 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562910-q8rvp"] Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.141713 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.144101 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562910-q8rvp"] Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.145289 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.145629 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.147027 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.208163 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2wgj\" (UniqueName: \"kubernetes.io/projected/0594a623-6ee6-4b42-beed-ef14ab5b2b72-kube-api-access-l2wgj\") pod \"auto-csr-approver-29562910-q8rvp\" (UID: \"0594a623-6ee6-4b42-beed-ef14ab5b2b72\") " pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.309129 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l2wgj\" (UniqueName: \"kubernetes.io/projected/0594a623-6ee6-4b42-beed-ef14ab5b2b72-kube-api-access-l2wgj\") pod \"auto-csr-approver-29562910-q8rvp\" (UID: \"0594a623-6ee6-4b42-beed-ef14ab5b2b72\") " pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.329128 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2wgj\" (UniqueName: \"kubernetes.io/projected/0594a623-6ee6-4b42-beed-ef14ab5b2b72-kube-api-access-l2wgj\") pod \"auto-csr-approver-29562910-q8rvp\" (UID: \"0594a623-6ee6-4b42-beed-ef14ab5b2b72\") " pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.463101 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:00 crc kubenswrapper[5110]: I0317 19:10:00.893954 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562910-q8rvp"] Mar 17 19:10:01 crc kubenswrapper[5110]: I0317 19:10:01.299490 5110 generic.go:358] "Generic (PLEG): container finished" podID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerID="5ff1bacb27c99674a6a50ec6cc1f83c5e2a2e3a0c1eaab88d9226377971cea6f" exitCode=0 Mar 17 19:10:01 crc kubenswrapper[5110]: I0317 19:10:01.299585 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerDied","Data":"5ff1bacb27c99674a6a50ec6cc1f83c5e2a2e3a0c1eaab88d9226377971cea6f"} Mar 17 19:10:01 crc kubenswrapper[5110]: I0317 19:10:01.301563 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" event={"ID":"0594a623-6ee6-4b42-beed-ef14ab5b2b72","Type":"ContainerStarted","Data":"508661a878a1699b79232f57166b795e7eed9c7aca347b4c6b59ee6b3defdc2e"} Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.550584 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645602 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-pull\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645659 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-proxy-ca-bundles\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645712 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildcachedir\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645782 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxtpn\" (UniqueName: \"kubernetes.io/projected/c9bfda01-07e3-4539-a8ed-857e07b64db3-kube-api-access-wxtpn\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645817 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-node-pullsecrets\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645859 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildworkdir\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645897 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-root\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645961 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-blob-cache\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.645984 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-push\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.646027 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-run\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.646103 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-system-configs\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.646201 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-ca-bundles\") pod \"c9bfda01-07e3-4539-a8ed-857e07b64db3\" (UID: \"c9bfda01-07e3-4539-a8ed-857e07b64db3\") " Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.646425 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.646462 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.646480 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.647326 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.647450 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.647870 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.649051 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.651100 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9bfda01-07e3-4539-a8ed-857e07b64db3-kube-api-access-wxtpn" (OuterVolumeSpecName: "kube-api-access-wxtpn") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "kube-api-access-wxtpn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.651203 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-push" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-push") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "builder-dockercfg-mpnmp-push". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.651218 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-pull" (OuterVolumeSpecName: "builder-dockercfg-mpnmp-pull") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "builder-dockercfg-mpnmp-pull". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748388 5110 reconciler_common.go:299] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748417 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-pull\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-pull\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748429 5110 reconciler_common.go:299] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748438 5110 reconciler_common.go:299] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748446 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wxtpn\" (UniqueName: \"kubernetes.io/projected/c9bfda01-07e3-4539-a8ed-857e07b64db3-kube-api-access-wxtpn\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748453 5110 reconciler_common.go:299] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9bfda01-07e3-4539-a8ed-857e07b64db3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748464 5110 reconciler_common.go:299] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748476 5110 reconciler_common.go:299] "Volume detached for volume \"builder-dockercfg-mpnmp-push\" (UniqueName: \"kubernetes.io/secret/c9bfda01-07e3-4539-a8ed-857e07b64db3-builder-dockercfg-mpnmp-push\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748487 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.748498 5110 reconciler_common.go:299] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.749827 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:10:02 crc kubenswrapper[5110]: I0317 19:10:02.849734 5110 reconciler_common.go:299] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.318974 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"c9bfda01-07e3-4539-a8ed-857e07b64db3","Type":"ContainerDied","Data":"3ffd32a68a1d873d28c7f4ec8ca7ccc7d4948ceef8f050b732b4be6fddb9eba6"} Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.319294 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ffd32a68a1d873d28c7f4ec8ca7ccc7d4948ceef8f050b732b4be6fddb9eba6" Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.319421 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.321447 5110 generic.go:358] "Generic (PLEG): container finished" podID="0594a623-6ee6-4b42-beed-ef14ab5b2b72" containerID="2ebdc7fb30206239dddc12c31f1ab0877d9e1a0f18ce9fc1dd435c91bcbd24ce" exitCode=0 Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.321637 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" event={"ID":"0594a623-6ee6-4b42-beed-ef14ab5b2b72","Type":"ContainerDied","Data":"2ebdc7fb30206239dddc12c31f1ab0877d9e1a0f18ce9fc1dd435c91bcbd24ce"} Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.580338 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "c9bfda01-07e3-4539-a8ed-857e07b64db3" (UID: "c9bfda01-07e3-4539-a8ed-857e07b64db3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:10:03 crc kubenswrapper[5110]: I0317 19:10:03.658796 5110 reconciler_common.go:299] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c9bfda01-07e3-4539-a8ed-857e07b64db3-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:04 crc kubenswrapper[5110]: I0317 19:10:04.555963 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:04 crc kubenswrapper[5110]: I0317 19:10:04.672764 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2wgj\" (UniqueName: \"kubernetes.io/projected/0594a623-6ee6-4b42-beed-ef14ab5b2b72-kube-api-access-l2wgj\") pod \"0594a623-6ee6-4b42-beed-ef14ab5b2b72\" (UID: \"0594a623-6ee6-4b42-beed-ef14ab5b2b72\") " Mar 17 19:10:04 crc kubenswrapper[5110]: I0317 19:10:04.682779 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0594a623-6ee6-4b42-beed-ef14ab5b2b72-kube-api-access-l2wgj" (OuterVolumeSpecName: "kube-api-access-l2wgj") pod "0594a623-6ee6-4b42-beed-ef14ab5b2b72" (UID: "0594a623-6ee6-4b42-beed-ef14ab5b2b72"). InnerVolumeSpecName "kube-api-access-l2wgj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:10:04 crc kubenswrapper[5110]: I0317 19:10:04.775918 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l2wgj\" (UniqueName: \"kubernetes.io/projected/0594a623-6ee6-4b42-beed-ef14ab5b2b72-kube-api-access-l2wgj\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:05 crc kubenswrapper[5110]: I0317 19:10:05.335215 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" event={"ID":"0594a623-6ee6-4b42-beed-ef14ab5b2b72","Type":"ContainerDied","Data":"508661a878a1699b79232f57166b795e7eed9c7aca347b4c6b59ee6b3defdc2e"} Mar 17 19:10:05 crc kubenswrapper[5110]: I0317 19:10:05.335262 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="508661a878a1699b79232f57166b795e7eed9c7aca347b4c6b59ee6b3defdc2e" Mar 17 19:10:05 crc kubenswrapper[5110]: I0317 19:10:05.335268 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562910-q8rvp" Mar 17 19:10:05 crc kubenswrapper[5110]: I0317 19:10:05.633963 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562904-wx68f"] Mar 17 19:10:05 crc kubenswrapper[5110]: I0317 19:10:05.644151 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562904-wx68f"] Mar 17 19:10:06 crc kubenswrapper[5110]: I0317 19:10:06.765552 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da3223b-917f-4267-bc6e-e5eb8c595f44" path="/var/lib/kubelet/pods/6da3223b-917f-4267-bc6e-e5eb8c595f44/volumes" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.877790 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-77448bb48c-q6znx"] Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878397 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0594a623-6ee6-4b42-beed-ef14ab5b2b72" containerName="oc" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878409 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="0594a623-6ee6-4b42-beed-ef14ab5b2b72" containerName="oc" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878423 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="git-clone" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878429 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="git-clone" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878443 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="manage-dockerfile" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878448 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="manage-dockerfile" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878462 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="docker-build" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878467 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="docker-build" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878552 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="c9bfda01-07e3-4539-a8ed-857e07b64db3" containerName="docker-build" Mar 17 19:10:07 crc kubenswrapper[5110]: I0317 19:10:07.878563 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="0594a623-6ee6-4b42-beed-ef14ab5b2b72" containerName="oc" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.191157 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-77448bb48c-q6znx"] Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.191341 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.199802 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-dockercfg-4vjhp\"" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.327085 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/a1d461c5-fb19-46b1-8791-ead59519666e-runner\") pod \"smart-gateway-operator-77448bb48c-q6znx\" (UID: \"a1d461c5-fb19-46b1-8791-ead59519666e\") " pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.327189 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtvt\" (UniqueName: \"kubernetes.io/projected/a1d461c5-fb19-46b1-8791-ead59519666e-kube-api-access-qxtvt\") pod \"smart-gateway-operator-77448bb48c-q6znx\" (UID: \"a1d461c5-fb19-46b1-8791-ead59519666e\") " pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.428980 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/a1d461c5-fb19-46b1-8791-ead59519666e-runner\") pod \"smart-gateway-operator-77448bb48c-q6znx\" (UID: \"a1d461c5-fb19-46b1-8791-ead59519666e\") " pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.429089 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qxtvt\" (UniqueName: \"kubernetes.io/projected/a1d461c5-fb19-46b1-8791-ead59519666e-kube-api-access-qxtvt\") pod \"smart-gateway-operator-77448bb48c-q6znx\" (UID: \"a1d461c5-fb19-46b1-8791-ead59519666e\") " pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.429802 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/a1d461c5-fb19-46b1-8791-ead59519666e-runner\") pod \"smart-gateway-operator-77448bb48c-q6znx\" (UID: \"a1d461c5-fb19-46b1-8791-ead59519666e\") " pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.447853 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxtvt\" (UniqueName: \"kubernetes.io/projected/a1d461c5-fb19-46b1-8791-ead59519666e-kube-api-access-qxtvt\") pod \"smart-gateway-operator-77448bb48c-q6znx\" (UID: \"a1d461c5-fb19-46b1-8791-ead59519666e\") " pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.509708 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" Mar 17 19:10:08 crc kubenswrapper[5110]: I0317 19:10:08.936507 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-77448bb48c-q6znx"] Mar 17 19:10:09 crc kubenswrapper[5110]: I0317 19:10:09.361131 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" event={"ID":"a1d461c5-fb19-46b1-8791-ead59519666e","Type":"ContainerStarted","Data":"3aaec10b6ba5106316b6ed35ddc228fe147c440a14fdd9d57e9864877d70366e"} Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.224194 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs"] Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.249989 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.255517 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs"] Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.257631 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-dockercfg-hflxp\"" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.378627 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/0c2d87e6-435e-4b36-aee5-e1c92a05821d-runner\") pod \"service-telemetry-operator-7d5d86f96d-jwhgs\" (UID: \"0c2d87e6-435e-4b36-aee5-e1c92a05821d\") " pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.378691 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhg8k\" (UniqueName: \"kubernetes.io/projected/0c2d87e6-435e-4b36-aee5-e1c92a05821d-kube-api-access-vhg8k\") pod \"service-telemetry-operator-7d5d86f96d-jwhgs\" (UID: \"0c2d87e6-435e-4b36-aee5-e1c92a05821d\") " pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.481442 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/0c2d87e6-435e-4b36-aee5-e1c92a05821d-runner\") pod \"service-telemetry-operator-7d5d86f96d-jwhgs\" (UID: \"0c2d87e6-435e-4b36-aee5-e1c92a05821d\") " pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.481921 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/0c2d87e6-435e-4b36-aee5-e1c92a05821d-runner\") pod \"service-telemetry-operator-7d5d86f96d-jwhgs\" (UID: \"0c2d87e6-435e-4b36-aee5-e1c92a05821d\") " pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.482021 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vhg8k\" (UniqueName: \"kubernetes.io/projected/0c2d87e6-435e-4b36-aee5-e1c92a05821d-kube-api-access-vhg8k\") pod \"service-telemetry-operator-7d5d86f96d-jwhgs\" (UID: \"0c2d87e6-435e-4b36-aee5-e1c92a05821d\") " pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.508711 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhg8k\" (UniqueName: \"kubernetes.io/projected/0c2d87e6-435e-4b36-aee5-e1c92a05821d-kube-api-access-vhg8k\") pod \"service-telemetry-operator-7d5d86f96d-jwhgs\" (UID: \"0c2d87e6-435e-4b36-aee5-e1c92a05821d\") " pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.578622 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" Mar 17 19:10:11 crc kubenswrapper[5110]: I0317 19:10:11.854209 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs"] Mar 17 19:10:11 crc kubenswrapper[5110]: W0317 19:10:11.862000 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c2d87e6_435e_4b36_aee5_e1c92a05821d.slice/crio-c55cf230491d7e5237305adf4eaf13f0f9dbc2936e1ed8faa2061eebda26febd WatchSource:0}: Error finding container c55cf230491d7e5237305adf4eaf13f0f9dbc2936e1ed8faa2061eebda26febd: Status 404 returned error can't find the container with id c55cf230491d7e5237305adf4eaf13f0f9dbc2936e1ed8faa2061eebda26febd Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.132975 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.133084 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.133142 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.133805 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb0b6120b2825b80ee94ecacb872639139788855e7496255ef0eb30ec1f502af"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.133874 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://cb0b6120b2825b80ee94ecacb872639139788855e7496255ef0eb30ec1f502af" gracePeriod=600 Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.387576 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" event={"ID":"0c2d87e6-435e-4b36-aee5-e1c92a05821d","Type":"ContainerStarted","Data":"c55cf230491d7e5237305adf4eaf13f0f9dbc2936e1ed8faa2061eebda26febd"} Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.390620 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="cb0b6120b2825b80ee94ecacb872639139788855e7496255ef0eb30ec1f502af" exitCode=0 Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.390669 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"cb0b6120b2825b80ee94ecacb872639139788855e7496255ef0eb30ec1f502af"} Mar 17 19:10:12 crc kubenswrapper[5110]: I0317 19:10:12.390959 5110 scope.go:117] "RemoveContainer" containerID="0938f11ace42511813dad951668253bd6932f3e91eca29ddfdba80f92fcd4637" Mar 17 19:10:21 crc kubenswrapper[5110]: I0317 19:10:21.482246 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerStarted","Data":"654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2"} Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.087660 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tvgt9"] Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.096291 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.112682 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvgt9"] Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.176805 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qs8s\" (UniqueName: \"kubernetes.io/projected/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-kube-api-access-2qs8s\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.176922 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-utilities\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.176952 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-catalog-content\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.278447 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2qs8s\" (UniqueName: \"kubernetes.io/projected/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-kube-api-access-2qs8s\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.278540 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-utilities\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.278573 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-catalog-content\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.279669 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-utilities\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.279790 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-catalog-content\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.311314 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qs8s\" (UniqueName: \"kubernetes.io/projected/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-kube-api-access-2qs8s\") pod \"community-operators-tvgt9\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.422236 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.514912 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" event={"ID":"a1d461c5-fb19-46b1-8791-ead59519666e","Type":"ContainerStarted","Data":"83a234e6e1be6a24ad3f1cb08fb702a8ea848f3f0dca05d0281da0af7fc5a382"} Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.532702 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-77448bb48c-q6znx" podStartSLOduration=2.556711978 podStartE2EDuration="18.532683948s" podCreationTimestamp="2026-03-17 19:10:07 +0000 UTC" firstStartedPulling="2026-03-17 19:10:08.942402268 +0000 UTC m=+1352.975013780" lastFinishedPulling="2026-03-17 19:10:24.918374228 +0000 UTC m=+1368.950985750" observedRunningTime="2026-03-17 19:10:25.530376565 +0000 UTC m=+1369.562988107" watchObservedRunningTime="2026-03-17 19:10:25.532683948 +0000 UTC m=+1369.565295470" Mar 17 19:10:25 crc kubenswrapper[5110]: I0317 19:10:25.968607 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvgt9"] Mar 17 19:10:26 crc kubenswrapper[5110]: I0317 19:10:26.521936 5110 generic.go:358] "Generic (PLEG): container finished" podID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerID="55af96cbb3b81d354546962cebf5cf4e55fee9511ffd12dc6d8a5015d162796d" exitCode=0 Mar 17 19:10:26 crc kubenswrapper[5110]: I0317 19:10:26.523342 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerDied","Data":"55af96cbb3b81d354546962cebf5cf4e55fee9511ffd12dc6d8a5015d162796d"} Mar 17 19:10:26 crc kubenswrapper[5110]: I0317 19:10:26.523365 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerStarted","Data":"8c7cf5ae241c7131e4929890a12807f05c5ee529db0e25aa11b7149c45740e42"} Mar 17 19:10:30 crc kubenswrapper[5110]: I0317 19:10:30.559116 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerStarted","Data":"7a32bbe16dbe30d16107eef5ad277614ea9f6c8ca1654c86f903ec96a5b60940"} Mar 17 19:10:30 crc kubenswrapper[5110]: I0317 19:10:30.561971 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" event={"ID":"0c2d87e6-435e-4b36-aee5-e1c92a05821d","Type":"ContainerStarted","Data":"dd7e0be45976320304027abbf41e8e39e4409b3274c247900d4fac1581610f56"} Mar 17 19:10:30 crc kubenswrapper[5110]: I0317 19:10:30.598072 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-7d5d86f96d-jwhgs" podStartSLOduration=1.23673246 podStartE2EDuration="19.598040318s" podCreationTimestamp="2026-03-17 19:10:11 +0000 UTC" firstStartedPulling="2026-03-17 19:10:11.864567708 +0000 UTC m=+1355.897179220" lastFinishedPulling="2026-03-17 19:10:30.225875556 +0000 UTC m=+1374.258487078" observedRunningTime="2026-03-17 19:10:30.595398166 +0000 UTC m=+1374.628009688" watchObservedRunningTime="2026-03-17 19:10:30.598040318 +0000 UTC m=+1374.630651840" Mar 17 19:10:31 crc kubenswrapper[5110]: I0317 19:10:31.570363 5110 generic.go:358] "Generic (PLEG): container finished" podID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerID="7a32bbe16dbe30d16107eef5ad277614ea9f6c8ca1654c86f903ec96a5b60940" exitCode=0 Mar 17 19:10:31 crc kubenswrapper[5110]: I0317 19:10:31.570431 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerDied","Data":"7a32bbe16dbe30d16107eef5ad277614ea9f6c8ca1654c86f903ec96a5b60940"} Mar 17 19:10:32 crc kubenswrapper[5110]: I0317 19:10:32.579987 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerStarted","Data":"40fb9bc93a75a5f93dd7099ba88e9de1ff8f7e8b21d22c1fd77d40324a345a94"} Mar 17 19:10:32 crc kubenswrapper[5110]: I0317 19:10:32.598875 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tvgt9" podStartSLOduration=3.919704103 podStartE2EDuration="7.598853989s" podCreationTimestamp="2026-03-17 19:10:25 +0000 UTC" firstStartedPulling="2026-03-17 19:10:26.52322335 +0000 UTC m=+1370.555834872" lastFinishedPulling="2026-03-17 19:10:30.202373236 +0000 UTC m=+1374.234984758" observedRunningTime="2026-03-17 19:10:32.596528216 +0000 UTC m=+1376.629139748" watchObservedRunningTime="2026-03-17 19:10:32.598853989 +0000 UTC m=+1376.631465511" Mar 17 19:10:35 crc kubenswrapper[5110]: I0317 19:10:35.423668 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:35 crc kubenswrapper[5110]: I0317 19:10:35.424587 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:35 crc kubenswrapper[5110]: I0317 19:10:35.480802 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:45 crc kubenswrapper[5110]: I0317 19:10:45.677824 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:45 crc kubenswrapper[5110]: I0317 19:10:45.722977 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvgt9"] Mar 17 19:10:46 crc kubenswrapper[5110]: I0317 19:10:46.686222 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tvgt9" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="registry-server" containerID="cri-o://40fb9bc93a75a5f93dd7099ba88e9de1ff8f7e8b21d22c1fd77d40324a345a94" gracePeriod=2 Mar 17 19:10:47 crc kubenswrapper[5110]: I0317 19:10:47.695650 5110 generic.go:358] "Generic (PLEG): container finished" podID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerID="40fb9bc93a75a5f93dd7099ba88e9de1ff8f7e8b21d22c1fd77d40324a345a94" exitCode=0 Mar 17 19:10:47 crc kubenswrapper[5110]: I0317 19:10:47.695718 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerDied","Data":"40fb9bc93a75a5f93dd7099ba88e9de1ff8f7e8b21d22c1fd77d40324a345a94"} Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.658456 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.714684 5110 scope.go:117] "RemoveContainer" containerID="1c46df711ac7b2cb5092328c8002dce4e55b3585a669047b9eb3fba4bd78da4b" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.718503 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvgt9" event={"ID":"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa","Type":"ContainerDied","Data":"8c7cf5ae241c7131e4929890a12807f05c5ee529db0e25aa11b7149c45740e42"} Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.718596 5110 scope.go:117] "RemoveContainer" containerID="40fb9bc93a75a5f93dd7099ba88e9de1ff8f7e8b21d22c1fd77d40324a345a94" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.718520 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvgt9" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.721596 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qs8s\" (UniqueName: \"kubernetes.io/projected/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-kube-api-access-2qs8s\") pod \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.721911 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-utilities\") pod \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.722002 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-catalog-content\") pod \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\" (UID: \"985cb8b8-afb1-43c7-8e68-c5a9e5e10efa\") " Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.723420 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-utilities" (OuterVolumeSpecName: "utilities") pod "985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" (UID: "985cb8b8-afb1-43c7-8e68-c5a9e5e10efa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.728450 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-kube-api-access-2qs8s" (OuterVolumeSpecName: "kube-api-access-2qs8s") pod "985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" (UID: "985cb8b8-afb1-43c7-8e68-c5a9e5e10efa"). InnerVolumeSpecName "kube-api-access-2qs8s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.792815 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" (UID: "985cb8b8-afb1-43c7-8e68-c5a9e5e10efa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.810287 5110 scope.go:117] "RemoveContainer" containerID="7a32bbe16dbe30d16107eef5ad277614ea9f6c8ca1654c86f903ec96a5b60940" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.823457 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2qs8s\" (UniqueName: \"kubernetes.io/projected/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-kube-api-access-2qs8s\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.823487 5110 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.823498 5110 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 19:10:48 crc kubenswrapper[5110]: I0317 19:10:48.837274 5110 scope.go:117] "RemoveContainer" containerID="55af96cbb3b81d354546962cebf5cf4e55fee9511ffd12dc6d8a5015d162796d" Mar 17 19:10:49 crc kubenswrapper[5110]: I0317 19:10:49.049532 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvgt9"] Mar 17 19:10:49 crc kubenswrapper[5110]: I0317 19:10:49.055221 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tvgt9"] Mar 17 19:10:50 crc kubenswrapper[5110]: I0317 19:10:50.753973 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" path="/var/lib/kubelet/pods/985cb8b8-afb1-43c7-8e68-c5a9e5e10efa/volumes" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.539923 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-qm4tl"] Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.540797 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="extract-content" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.540817 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="extract-content" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.540836 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="extract-utilities" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.540844 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="extract-utilities" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.540868 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="registry-server" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.540876 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="registry-server" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.541007 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="985cb8b8-afb1-43c7-8e68-c5a9e5e10efa" containerName="registry-server" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.565835 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-qm4tl"] Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.566042 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.573023 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-openstack-ca\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.573282 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-dockercfg-gr5tx\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.573407 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-openstack-credentials\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.575241 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-inter-router-credentials\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.575761 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-interconnect-sasl-config\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.576713 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-users\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.576737 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-inter-router-ca\"" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679300 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679558 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679580 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-users\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679625 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6hzw\" (UniqueName: \"kubernetes.io/projected/b47bd39b-545a-4b18-9a09-e85804e73449-kube-api-access-b6hzw\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679646 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-config\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679784 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.679928 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781166 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b6hzw\" (UniqueName: \"kubernetes.io/projected/b47bd39b-545a-4b18-9a09-e85804e73449-kube-api-access-b6hzw\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781226 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-config\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781256 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781298 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781329 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781346 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.781452 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-users\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.782367 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-config\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.788325 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.788610 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.789024 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.795627 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.799166 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-users\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.800605 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6hzw\" (UniqueName: \"kubernetes.io/projected/b47bd39b-545a-4b18-9a09-e85804e73449-kube-api-access-b6hzw\") pod \"default-interconnect-55bf8d5cb-qm4tl\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:52 crc kubenswrapper[5110]: I0317 19:10:52.891047 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:10:53 crc kubenswrapper[5110]: I0317 19:10:53.140404 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-qm4tl"] Mar 17 19:10:53 crc kubenswrapper[5110]: I0317 19:10:53.758074 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" event={"ID":"b47bd39b-545a-4b18-9a09-e85804e73449","Type":"ContainerStarted","Data":"625f6492479425830135ed971fd7e95fc96b1a1a11199141ea843156ad0992f8"} Mar 17 19:10:57 crc kubenswrapper[5110]: I0317 19:10:57.799901 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" event={"ID":"b47bd39b-545a-4b18-9a09-e85804e73449","Type":"ContainerStarted","Data":"19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d"} Mar 17 19:10:57 crc kubenswrapper[5110]: I0317 19:10:57.822929 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" podStartSLOduration=1.42795681 podStartE2EDuration="5.822904431s" podCreationTimestamp="2026-03-17 19:10:52 +0000 UTC" firstStartedPulling="2026-03-17 19:10:53.144021763 +0000 UTC m=+1397.176633295" lastFinishedPulling="2026-03-17 19:10:57.538969394 +0000 UTC m=+1401.571580916" observedRunningTime="2026-03-17 19:10:57.81700804 +0000 UTC m=+1401.849619592" watchObservedRunningTime="2026-03-17 19:10:57.822904431 +0000 UTC m=+1401.855515963" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.659328 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.669106 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.671710 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-session-secret\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.672129 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-prometheus-proxy-tls\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.672731 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"serving-certs-ca-bundle\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.672892 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-2\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.672739 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default-tls-assets-0\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.673131 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-0\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.673907 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-stf-dockercfg-5fb2t\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.674396 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-1\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.674991 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.677500 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default-web-config\"" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.690897 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.828116 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p24fk\" (UniqueName: \"kubernetes.io/projected/978872ef-66de-48f8-b24d-ae6176a05c08-kube-api-access-p24fk\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.828551 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.828670 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.828769 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/978872ef-66de-48f8-b24d-ae6176a05c08-tls-assets\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.828852 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/978872ef-66de-48f8-b24d-ae6176a05c08-config-out\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.828936 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.829022 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.829120 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.829209 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-config\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.829303 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.829407 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-web-config\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.829570 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.930839 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.930937 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.930990 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/978872ef-66de-48f8-b24d-ae6176a05c08-tls-assets\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931042 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/978872ef-66de-48f8-b24d-ae6176a05c08-config-out\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931097 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931163 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931188 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931241 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-config\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931343 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931432 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-web-config\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931504 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.931548 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p24fk\" (UniqueName: \"kubernetes.io/projected/978872ef-66de-48f8-b24d-ae6176a05c08-kube-api-access-p24fk\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: E0317 19:11:02.931773 5110 secret.go:189] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 17 19:11:02 crc kubenswrapper[5110]: E0317 19:11:02.931962 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls podName:978872ef-66de-48f8-b24d-ae6176a05c08 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:03.431934852 +0000 UTC m=+1407.464546424 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "978872ef-66de-48f8-b24d-ae6176a05c08") : secret "default-prometheus-proxy-tls" not found Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.932500 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.932967 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.933232 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.934288 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/978872ef-66de-48f8-b24d-ae6176a05c08-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.937572 5110 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.937620 5110 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/24c950925a52d0fe7b2315a596f884d6c28af74ce1adad0879c4b212d1753d43/globalmount\"" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.938407 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/978872ef-66de-48f8-b24d-ae6176a05c08-config-out\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.940160 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.941353 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-web-config\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.941495 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-config\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.941868 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/978872ef-66de-48f8-b24d-ae6176a05c08-tls-assets\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.952919 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p24fk\" (UniqueName: \"kubernetes.io/projected/978872ef-66de-48f8-b24d-ae6176a05c08-kube-api-access-p24fk\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:02 crc kubenswrapper[5110]: I0317 19:11:02.970868 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a34f90d5-8e17-445c-b66a-8911e17d93dc\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:03 crc kubenswrapper[5110]: I0317 19:11:03.440117 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:03 crc kubenswrapper[5110]: E0317 19:11:03.440325 5110 secret.go:189] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 17 19:11:03 crc kubenswrapper[5110]: E0317 19:11:03.440403 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls podName:978872ef-66de-48f8-b24d-ae6176a05c08 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:04.440385177 +0000 UTC m=+1408.472996699 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "978872ef-66de-48f8-b24d-ae6176a05c08") : secret "default-prometheus-proxy-tls" not found Mar 17 19:11:04 crc kubenswrapper[5110]: I0317 19:11:04.454074 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:04 crc kubenswrapper[5110]: I0317 19:11:04.461762 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/978872ef-66de-48f8-b24d-ae6176a05c08-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"978872ef-66de-48f8-b24d-ae6176a05c08\") " pod="service-telemetry/prometheus-default-0" Mar 17 19:11:04 crc kubenswrapper[5110]: I0317 19:11:04.487815 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:04 crc kubenswrapper[5110]: I0317 19:11:04.760297 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 17 19:11:04 crc kubenswrapper[5110]: I0317 19:11:04.853526 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"978872ef-66de-48f8-b24d-ae6176a05c08","Type":"ContainerStarted","Data":"fa5b83887acf10b45199243a4e52e2ba2ab51d543484c8abea9b50ef41765ddc"} Mar 17 19:11:09 crc kubenswrapper[5110]: I0317 19:11:09.896561 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"978872ef-66de-48f8-b24d-ae6176a05c08","Type":"ContainerStarted","Data":"688d08516dd66d2a1cbc80537270d76f8b1fbf3bb42239fdd8ffa0f6ea5c7179"} Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.067991 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-694dc457d5-rcxnv"] Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.077924 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-694dc457d5-rcxnv"] Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.078018 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.174369 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r28qj\" (UniqueName: \"kubernetes.io/projected/4a47a709-2c00-40b6-bd44-90ee935a14f5-kube-api-access-r28qj\") pod \"default-snmp-webhook-694dc457d5-rcxnv\" (UID: \"4a47a709-2c00-40b6-bd44-90ee935a14f5\") " pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.276228 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r28qj\" (UniqueName: \"kubernetes.io/projected/4a47a709-2c00-40b6-bd44-90ee935a14f5-kube-api-access-r28qj\") pod \"default-snmp-webhook-694dc457d5-rcxnv\" (UID: \"4a47a709-2c00-40b6-bd44-90ee935a14f5\") " pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.301384 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r28qj\" (UniqueName: \"kubernetes.io/projected/4a47a709-2c00-40b6-bd44-90ee935a14f5-kube-api-access-r28qj\") pod \"default-snmp-webhook-694dc457d5-rcxnv\" (UID: \"4a47a709-2c00-40b6-bd44-90ee935a14f5\") " pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.394181 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.820555 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-694dc457d5-rcxnv"] Mar 17 19:11:13 crc kubenswrapper[5110]: W0317 19:11:13.825760 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a47a709_2c00_40b6_bd44_90ee935a14f5.slice/crio-7d279bbc818b8d6d470f264c759a349f10705d995bb39254384ef5b0288184de WatchSource:0}: Error finding container 7d279bbc818b8d6d470f264c759a349f10705d995bb39254384ef5b0288184de: Status 404 returned error can't find the container with id 7d279bbc818b8d6d470f264c759a349f10705d995bb39254384ef5b0288184de Mar 17 19:11:13 crc kubenswrapper[5110]: I0317 19:11:13.930135 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" event={"ID":"4a47a709-2c00-40b6-bd44-90ee935a14f5","Type":"ContainerStarted","Data":"7d279bbc818b8d6d470f264c759a349f10705d995bb39254384ef5b0288184de"} Mar 17 19:11:14 crc kubenswrapper[5110]: I0317 19:11:14.942593 5110 generic.go:358] "Generic (PLEG): container finished" podID="978872ef-66de-48f8-b24d-ae6176a05c08" containerID="688d08516dd66d2a1cbc80537270d76f8b1fbf3bb42239fdd8ffa0f6ea5c7179" exitCode=0 Mar 17 19:11:14 crc kubenswrapper[5110]: I0317 19:11:14.942683 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"978872ef-66de-48f8-b24d-ae6176a05c08","Type":"ContainerDied","Data":"688d08516dd66d2a1cbc80537270d76f8b1fbf3bb42239fdd8ffa0f6ea5c7179"} Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.920436 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.972664 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.972845 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.975183 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-cluster-tls-config\"" Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.975579 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-tls-assets-0\"" Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.975904 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-generated\"" Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.976105 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-stf-dockercfg-g7qd7\"" Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.976277 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-web-config\"" Mar 17 19:11:16 crc kubenswrapper[5110]: I0317 19:11:16.979569 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-alertmanager-proxy-tls\"" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026451 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpvb4\" (UniqueName: \"kubernetes.io/projected/da51ef37-5384-4ba3-87d8-7b1feadac8dc-kube-api-access-dpvb4\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026557 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-web-config\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026593 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/da51ef37-5384-4ba3-87d8-7b1feadac8dc-tls-assets\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026625 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/da51ef37-5384-4ba3-87d8-7b1feadac8dc-config-out\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026739 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026808 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026882 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026900 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.026934 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-config-volume\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.128081 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-web-config\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.128161 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/da51ef37-5384-4ba3-87d8-7b1feadac8dc-tls-assets\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.128195 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/da51ef37-5384-4ba3-87d8-7b1feadac8dc-config-out\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.128239 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.128287 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: E0317 19:11:17.128496 5110 secret.go:189] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 17 19:11:17 crc kubenswrapper[5110]: E0317 19:11:17.128657 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls podName:da51ef37-5384-4ba3-87d8-7b1feadac8dc nodeName:}" failed. No retries permitted until 2026-03-17 19:11:17.628622817 +0000 UTC m=+1421.661234339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "da51ef37-5384-4ba3-87d8-7b1feadac8dc") : secret "default-alertmanager-proxy-tls" not found Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.130971 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.131065 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.131139 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-config-volume\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.131242 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dpvb4\" (UniqueName: \"kubernetes.io/projected/da51ef37-5384-4ba3-87d8-7b1feadac8dc-kube-api-access-dpvb4\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.135281 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.139346 5110 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.139567 5110 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7305cee48a54a4c36fb66b93e89aa470b6abc41c46f41ad7a28849349b9c6d44/globalmount\"" pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.141855 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.141967 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/da51ef37-5384-4ba3-87d8-7b1feadac8dc-tls-assets\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.144152 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-config-volume\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.155356 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-web-config\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.155637 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpvb4\" (UniqueName: \"kubernetes.io/projected/da51ef37-5384-4ba3-87d8-7b1feadac8dc-kube-api-access-dpvb4\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.163594 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/da51ef37-5384-4ba3-87d8-7b1feadac8dc-config-out\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.170573 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13d6bad5-d323-48d6-bea1-ed6c3da61b49\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: I0317 19:11:17.639498 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:17 crc kubenswrapper[5110]: E0317 19:11:17.639674 5110 secret.go:189] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 17 19:11:17 crc kubenswrapper[5110]: E0317 19:11:17.639734 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls podName:da51ef37-5384-4ba3-87d8-7b1feadac8dc nodeName:}" failed. No retries permitted until 2026-03-17 19:11:18.639718605 +0000 UTC m=+1422.672330127 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "da51ef37-5384-4ba3-87d8-7b1feadac8dc") : secret "default-alertmanager-proxy-tls" not found Mar 17 19:11:18 crc kubenswrapper[5110]: I0317 19:11:18.654604 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:18 crc kubenswrapper[5110]: E0317 19:11:18.654812 5110 secret.go:189] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 17 19:11:18 crc kubenswrapper[5110]: E0317 19:11:18.655256 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls podName:da51ef37-5384-4ba3-87d8-7b1feadac8dc nodeName:}" failed. No retries permitted until 2026-03-17 19:11:20.655048422 +0000 UTC m=+1424.687659944 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "da51ef37-5384-4ba3-87d8-7b1feadac8dc") : secret "default-alertmanager-proxy-tls" not found Mar 17 19:11:20 crc kubenswrapper[5110]: I0317 19:11:20.683081 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:20 crc kubenswrapper[5110]: I0317 19:11:20.709323 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/da51ef37-5384-4ba3-87d8-7b1feadac8dc-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"da51ef37-5384-4ba3-87d8-7b1feadac8dc\") " pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:20 crc kubenswrapper[5110]: I0317 19:11:20.901232 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 17 19:11:21 crc kubenswrapper[5110]: I0317 19:11:21.947377 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 17 19:11:21 crc kubenswrapper[5110]: I0317 19:11:21.989091 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" event={"ID":"4a47a709-2c00-40b6-bd44-90ee935a14f5","Type":"ContainerStarted","Data":"bc1f31f6e17c66cb77766c9c7631928b68142afb0c782c30ae7c133681bf98e4"} Mar 17 19:11:22 crc kubenswrapper[5110]: I0317 19:11:22.011541 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-694dc457d5-rcxnv" podStartSLOduration=1.202465269 podStartE2EDuration="9.011521115s" podCreationTimestamp="2026-03-17 19:11:13 +0000 UTC" firstStartedPulling="2026-03-17 19:11:13.828413197 +0000 UTC m=+1417.861024719" lastFinishedPulling="2026-03-17 19:11:21.637469043 +0000 UTC m=+1425.670080565" observedRunningTime="2026-03-17 19:11:22.003609159 +0000 UTC m=+1426.036220681" watchObservedRunningTime="2026-03-17 19:11:22.011521115 +0000 UTC m=+1426.044132637" Mar 17 19:11:22 crc kubenswrapper[5110]: W0317 19:11:22.143120 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda51ef37_5384_4ba3_87d8_7b1feadac8dc.slice/crio-3496214a99b6583748ccd46701c40054cae1ec35b3b309d0cadde10b969cdbc7 WatchSource:0}: Error finding container 3496214a99b6583748ccd46701c40054cae1ec35b3b309d0cadde10b969cdbc7: Status 404 returned error can't find the container with id 3496214a99b6583748ccd46701c40054cae1ec35b3b309d0cadde10b969cdbc7 Mar 17 19:11:22 crc kubenswrapper[5110]: I0317 19:11:22.997327 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"da51ef37-5384-4ba3-87d8-7b1feadac8dc","Type":"ContainerStarted","Data":"3496214a99b6583748ccd46701c40054cae1ec35b3b309d0cadde10b969cdbc7"} Mar 17 19:11:24 crc kubenswrapper[5110]: I0317 19:11:24.006420 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"da51ef37-5384-4ba3-87d8-7b1feadac8dc","Type":"ContainerStarted","Data":"2cb158501c8860fd485ca88c5dc206aa374dbf67283195507ba3f1ab1609fb0c"} Mar 17 19:11:27 crc kubenswrapper[5110]: I0317 19:11:27.028806 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"978872ef-66de-48f8-b24d-ae6176a05c08","Type":"ContainerStarted","Data":"f33151b611df6f1a225837f0a7d3809c0ac72ad97dc32674ae1eccc682001af1"} Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.051813 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"978872ef-66de-48f8-b24d-ae6176a05c08","Type":"ContainerStarted","Data":"a38ae147d626055f31d78cb1667798d0c2a375351d1bd17318adcb6bb9a6df87"} Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.053646 5110 generic.go:358] "Generic (PLEG): container finished" podID="da51ef37-5384-4ba3-87d8-7b1feadac8dc" containerID="2cb158501c8860fd485ca88c5dc206aa374dbf67283195507ba3f1ab1609fb0c" exitCode=0 Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.053672 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"da51ef37-5384-4ba3-87d8-7b1feadac8dc","Type":"ContainerDied","Data":"2cb158501c8860fd485ca88c5dc206aa374dbf67283195507ba3f1ab1609fb0c"} Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.081046 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh"] Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.086273 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.090013 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-session-secret\"" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.090212 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-meter-proxy-tls\"" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.090290 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-meter-sg-core-configmap\"" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.090213 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-dockercfg-n58jh\"" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.093310 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh"] Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.240749 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghn6\" (UniqueName: \"kubernetes.io/projected/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-kube-api-access-xghn6\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.240911 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.240951 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.241006 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.241081 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.342865 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xghn6\" (UniqueName: \"kubernetes.io/projected/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-kube-api-access-xghn6\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.343260 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.343299 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.343352 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.343399 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: E0317 19:11:30.343675 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 17 19:11:30 crc kubenswrapper[5110]: E0317 19:11:30.343761 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls podName:e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad nodeName:}" failed. No retries permitted until 2026-03-17 19:11:30.843731737 +0000 UTC m=+1434.876343259 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" (UID: "e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.343924 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.344047 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.362567 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.363026 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghn6\" (UniqueName: \"kubernetes.io/projected/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-kube-api-access-xghn6\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: I0317 19:11:30.880807 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:30 crc kubenswrapper[5110]: E0317 19:11:30.881899 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 17 19:11:30 crc kubenswrapper[5110]: E0317 19:11:30.881979 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls podName:e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad nodeName:}" failed. No retries permitted until 2026-03-17 19:11:31.881959234 +0000 UTC m=+1435.914570756 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" (UID: "e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 17 19:11:31 crc kubenswrapper[5110]: I0317 19:11:31.901919 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:31 crc kubenswrapper[5110]: I0317 19:11:31.912601 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh\" (UID: \"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:32 crc kubenswrapper[5110]: I0317 19:11:32.202510 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" Mar 17 19:11:32 crc kubenswrapper[5110]: I0317 19:11:32.791156 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf"] Mar 17 19:11:32 crc kubenswrapper[5110]: I0317 19:11:32.863806 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf"] Mar 17 19:11:32 crc kubenswrapper[5110]: I0317 19:11:32.863968 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:32 crc kubenswrapper[5110]: I0317 19:11:32.866345 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-meter-sg-core-configmap\"" Mar 17 19:11:32 crc kubenswrapper[5110]: I0317 19:11:32.866606 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-meter-proxy-tls\"" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.019705 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ddba8057-d741-42e2-8e66-d1a5f0ed2524-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.020197 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ddba8057-d741-42e2-8e66-d1a5f0ed2524-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.020256 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.022180 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmw5p\" (UniqueName: \"kubernetes.io/projected/ddba8057-d741-42e2-8e66-d1a5f0ed2524-kube-api-access-mmw5p\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.022518 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.124008 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.124093 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ddba8057-d741-42e2-8e66-d1a5f0ed2524-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.124113 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ddba8057-d741-42e2-8e66-d1a5f0ed2524-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.124141 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.124326 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mmw5p\" (UniqueName: \"kubernetes.io/projected/ddba8057-d741-42e2-8e66-d1a5f0ed2524-kube-api-access-mmw5p\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: E0317 19:11:33.124414 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 17 19:11:33 crc kubenswrapper[5110]: E0317 19:11:33.124557 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls podName:ddba8057-d741-42e2-8e66-d1a5f0ed2524 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:33.624525763 +0000 UTC m=+1437.657137315 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" (UID: "ddba8057-d741-42e2-8e66-d1a5f0ed2524") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.124811 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ddba8057-d741-42e2-8e66-d1a5f0ed2524-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.125679 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ddba8057-d741-42e2-8e66-d1a5f0ed2524-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.136609 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.142650 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmw5p\" (UniqueName: \"kubernetes.io/projected/ddba8057-d741-42e2-8e66-d1a5f0ed2524-kube-api-access-mmw5p\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: I0317 19:11:33.631229 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:33 crc kubenswrapper[5110]: E0317 19:11:33.631711 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 17 19:11:33 crc kubenswrapper[5110]: E0317 19:11:33.631847 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls podName:ddba8057-d741-42e2-8e66-d1a5f0ed2524 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:34.631827637 +0000 UTC m=+1438.664439159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" (UID: "ddba8057-d741-42e2-8e66-d1a5f0ed2524") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 17 19:11:34 crc kubenswrapper[5110]: I0317 19:11:34.642600 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:34 crc kubenswrapper[5110]: I0317 19:11:34.648660 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba8057-d741-42e2-8e66-d1a5f0ed2524-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf\" (UID: \"ddba8057-d741-42e2-8e66-d1a5f0ed2524\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:34 crc kubenswrapper[5110]: I0317 19:11:34.680263 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" Mar 17 19:11:36 crc kubenswrapper[5110]: I0317 19:11:36.761907 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh"] Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.110017 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg"] Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.117983 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.121441 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-sens-meter-sg-core-configmap\"" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.121713 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-sens-meter-proxy-tls\"" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.126172 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg"] Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.173334 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/091754de-5b5e-438e-ab68-395e08cdccb3-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.173404 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftmqc\" (UniqueName: \"kubernetes.io/projected/091754de-5b5e-438e-ab68-395e08cdccb3-kube-api-access-ftmqc\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.173517 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.173733 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.173762 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/091754de-5b5e-438e-ab68-395e08cdccb3-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.275467 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.275511 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.275531 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/091754de-5b5e-438e-ab68-395e08cdccb3-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.275596 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/091754de-5b5e-438e-ab68-395e08cdccb3-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.275618 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ftmqc\" (UniqueName: \"kubernetes.io/projected/091754de-5b5e-438e-ab68-395e08cdccb3-kube-api-access-ftmqc\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: E0317 19:11:37.276034 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 17 19:11:37 crc kubenswrapper[5110]: E0317 19:11:37.276096 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls podName:091754de-5b5e-438e-ab68-395e08cdccb3 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:37.776080782 +0000 UTC m=+1441.808692304 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" (UID: "091754de-5b5e-438e-ab68-395e08cdccb3") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.276884 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/091754de-5b5e-438e-ab68-395e08cdccb3-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.277151 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/091754de-5b5e-438e-ab68-395e08cdccb3-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.288941 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.292137 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftmqc\" (UniqueName: \"kubernetes.io/projected/091754de-5b5e-438e-ab68-395e08cdccb3-kube-api-access-ftmqc\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: W0317 19:11:37.655427 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode32a7feb_ef4c_4949_8dec_5c2d3f2f48ad.slice/crio-0f4453d620acf5c60728a1fce7e76dda52597b52951abd73667a137fdc9446fd WatchSource:0}: Error finding container 0f4453d620acf5c60728a1fce7e76dda52597b52951abd73667a137fdc9446fd: Status 404 returned error can't find the container with id 0f4453d620acf5c60728a1fce7e76dda52597b52951abd73667a137fdc9446fd Mar 17 19:11:37 crc kubenswrapper[5110]: I0317 19:11:37.790865 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:37 crc kubenswrapper[5110]: E0317 19:11:37.791048 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 17 19:11:37 crc kubenswrapper[5110]: E0317 19:11:37.791285 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls podName:091754de-5b5e-438e-ab68-395e08cdccb3 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:38.791267041 +0000 UTC m=+1442.823878563 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" (UID: "091754de-5b5e-438e-ab68-395e08cdccb3") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 17 19:11:38 crc kubenswrapper[5110]: I0317 19:11:38.106157 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"da51ef37-5384-4ba3-87d8-7b1feadac8dc","Type":"ContainerStarted","Data":"076cd0cf28426a9869e78a7bf3b8dc8e472fca91bc4de3bb02ed4fbbe3f7aacb"} Mar 17 19:11:38 crc kubenswrapper[5110]: I0317 19:11:38.108329 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"978872ef-66de-48f8-b24d-ae6176a05c08","Type":"ContainerStarted","Data":"cd07c3755276a1d32dfa14c5dbee52357d8ce3c70605558eccdab6d520aa1b36"} Mar 17 19:11:38 crc kubenswrapper[5110]: I0317 19:11:38.110177 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerStarted","Data":"0f4453d620acf5c60728a1fce7e76dda52597b52951abd73667a137fdc9446fd"} Mar 17 19:11:38 crc kubenswrapper[5110]: I0317 19:11:38.135128 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf"] Mar 17 19:11:38 crc kubenswrapper[5110]: W0317 19:11:38.136400 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddba8057_d741_42e2_8e66_d1a5f0ed2524.slice/crio-2df0e6c32e16c7efd27c2f12e5998d10d992ef2e2ce08d1f5bacc1b4d916ad6d WatchSource:0}: Error finding container 2df0e6c32e16c7efd27c2f12e5998d10d992ef2e2ce08d1f5bacc1b4d916ad6d: Status 404 returned error can't find the container with id 2df0e6c32e16c7efd27c2f12e5998d10d992ef2e2ce08d1f5bacc1b4d916ad6d Mar 17 19:11:38 crc kubenswrapper[5110]: I0317 19:11:38.143324 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=4.168190818 podStartE2EDuration="37.143307624s" podCreationTimestamp="2026-03-17 19:11:01 +0000 UTC" firstStartedPulling="2026-03-17 19:11:04.751708919 +0000 UTC m=+1408.784320451" lastFinishedPulling="2026-03-17 19:11:37.726825735 +0000 UTC m=+1441.759437257" observedRunningTime="2026-03-17 19:11:38.131102212 +0000 UTC m=+1442.163713744" watchObservedRunningTime="2026-03-17 19:11:38.143307624 +0000 UTC m=+1442.175919146" Mar 17 19:11:38 crc kubenswrapper[5110]: I0317 19:11:38.810860 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:38 crc kubenswrapper[5110]: E0317 19:11:38.811014 5110 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 17 19:11:38 crc kubenswrapper[5110]: E0317 19:11:38.811280 5110 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls podName:091754de-5b5e-438e-ab68-395e08cdccb3 nodeName:}" failed. No retries permitted until 2026-03-17 19:11:40.811258236 +0000 UTC m=+1444.843869758 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" (UID: "091754de-5b5e-438e-ab68-395e08cdccb3") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 17 19:11:39 crc kubenswrapper[5110]: I0317 19:11:39.117526 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerStarted","Data":"2df0e6c32e16c7efd27c2f12e5998d10d992ef2e2ce08d1f5bacc1b4d916ad6d"} Mar 17 19:11:39 crc kubenswrapper[5110]: I0317 19:11:39.119084 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerStarted","Data":"41ee87ff507a4de0a40dc7b80f42711f0e6ea8ac0e84df1c650911a345adea40"} Mar 17 19:11:39 crc kubenswrapper[5110]: I0317 19:11:39.487999 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:40 crc kubenswrapper[5110]: I0317 19:11:40.127783 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerStarted","Data":"1e3df99559593b3cbeb74186607aaa2bce06c867a2454c2252cb36b19734502a"} Mar 17 19:11:40 crc kubenswrapper[5110]: I0317 19:11:40.127827 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerStarted","Data":"8174ece36f7e1e03bf9bb92c708bdcd311e1068346de130bfb11257318dbcbe2"} Mar 17 19:11:40 crc kubenswrapper[5110]: I0317 19:11:40.129861 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerStarted","Data":"77147d07f1f29426cd72ef9ab476deb9f2f2bc446b35adadc496222fa940ae56"} Mar 17 19:11:40 crc kubenswrapper[5110]: I0317 19:11:40.133874 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"da51ef37-5384-4ba3-87d8-7b1feadac8dc","Type":"ContainerStarted","Data":"f6d1554b3e892a5fda4d1fbc5718a574fdd70bce233c9640916c346125713348"} Mar 17 19:11:40 crc kubenswrapper[5110]: I0317 19:11:40.839244 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:40 crc kubenswrapper[5110]: I0317 19:11:40.849902 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/091754de-5b5e-438e-ab68-395e08cdccb3-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg\" (UID: \"091754de-5b5e-438e-ab68-395e08cdccb3\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:41 crc kubenswrapper[5110]: I0317 19:11:41.040828 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" Mar 17 19:11:41 crc kubenswrapper[5110]: I0317 19:11:41.146895 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"da51ef37-5384-4ba3-87d8-7b1feadac8dc","Type":"ContainerStarted","Data":"c1c3381f41beaeb39fb28b1eb249265b39b8644a0d927d3dd784e9c7d011c1b5"} Mar 17 19:11:41 crc kubenswrapper[5110]: I0317 19:11:41.192972 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=15.839352252 podStartE2EDuration="26.192950756s" podCreationTimestamp="2026-03-17 19:11:15 +0000 UTC" firstStartedPulling="2026-03-17 19:11:30.05466832 +0000 UTC m=+1434.087279842" lastFinishedPulling="2026-03-17 19:11:40.408266824 +0000 UTC m=+1444.440878346" observedRunningTime="2026-03-17 19:11:41.173694421 +0000 UTC m=+1445.206305963" watchObservedRunningTime="2026-03-17 19:11:41.192950756 +0000 UTC m=+1445.225562268" Mar 17 19:11:41 crc kubenswrapper[5110]: I0317 19:11:41.478449 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg"] Mar 17 19:11:42 crc kubenswrapper[5110]: I0317 19:11:42.165868 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerStarted","Data":"5e5e49aec60f83799f3f44869dfbf026f75a5b5aca95b81f7824596df494a382"} Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.480197 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s"] Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.515455 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s"] Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.515607 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.518503 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-cert\"" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.518651 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-event-sg-core-configmap\"" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.675149 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvznd\" (UniqueName: \"kubernetes.io/projected/9c8d3727-791f-4128-ad7b-05105eac0743-kube-api-access-qvznd\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.675449 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9c8d3727-791f-4128-ad7b-05105eac0743-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.675613 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/9c8d3727-791f-4128-ad7b-05105eac0743-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.675741 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/9c8d3727-791f-4128-ad7b-05105eac0743-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.777188 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9c8d3727-791f-4128-ad7b-05105eac0743-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.777570 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/9c8d3727-791f-4128-ad7b-05105eac0743-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.777614 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/9c8d3727-791f-4128-ad7b-05105eac0743-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.777692 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qvznd\" (UniqueName: \"kubernetes.io/projected/9c8d3727-791f-4128-ad7b-05105eac0743-kube-api-access-qvznd\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.777761 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9c8d3727-791f-4128-ad7b-05105eac0743-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.778712 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/9c8d3727-791f-4128-ad7b-05105eac0743-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.793728 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/9c8d3727-791f-4128-ad7b-05105eac0743-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.795797 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvznd\" (UniqueName: \"kubernetes.io/projected/9c8d3727-791f-4128-ad7b-05105eac0743-kube-api-access-qvznd\") pod \"default-cloud1-coll-event-smartgateway-59876f8548-vkd7s\" (UID: \"9c8d3727-791f-4128-ad7b-05105eac0743\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:43 crc kubenswrapper[5110]: I0317 19:11:43.851654 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.181511 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerStarted","Data":"594611d67ddec25cdfc97de6f2053015fbedb511aaf1af88b5b2999310d8dcdd"} Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.181819 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerStarted","Data":"31cd2b73b7ad916f59595243040588ed6a5cf305b4e38b4ccad1a81faee2bc8c"} Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.181830 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerStarted","Data":"c9dc6f484a2fe3eaa9f862e8602a229865ade60937e83f496bebf742eef81962"} Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.183309 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerStarted","Data":"df02ded3dc7b399c709ecd58e69ec22ae91b4f231725dfc151f34e4d16657fd2"} Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.184929 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerStarted","Data":"62681aae4ca2b6683364f392723b3e5510626cca72ecfa53cf94542b89250ebf"} Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.224711 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" podStartSLOduration=4.885724113 podStartE2EDuration="7.224689329s" podCreationTimestamp="2026-03-17 19:11:37 +0000 UTC" firstStartedPulling="2026-03-17 19:11:41.500369943 +0000 UTC m=+1445.532981485" lastFinishedPulling="2026-03-17 19:11:43.839335179 +0000 UTC m=+1447.871946701" observedRunningTime="2026-03-17 19:11:44.205595229 +0000 UTC m=+1448.238206751" watchObservedRunningTime="2026-03-17 19:11:44.224689329 +0000 UTC m=+1448.257300851" Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.224967 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" podStartSLOduration=7.042695732 podStartE2EDuration="12.224962177s" podCreationTimestamp="2026-03-17 19:11:32 +0000 UTC" firstStartedPulling="2026-03-17 19:11:38.138329819 +0000 UTC m=+1442.170941341" lastFinishedPulling="2026-03-17 19:11:43.320596264 +0000 UTC m=+1447.353207786" observedRunningTime="2026-03-17 19:11:44.224145245 +0000 UTC m=+1448.256756767" watchObservedRunningTime="2026-03-17 19:11:44.224962177 +0000 UTC m=+1448.257573699" Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.243478 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" podStartSLOduration=8.609690042 podStartE2EDuration="14.243456651s" podCreationTimestamp="2026-03-17 19:11:30 +0000 UTC" firstStartedPulling="2026-03-17 19:11:37.657772093 +0000 UTC m=+1441.690383615" lastFinishedPulling="2026-03-17 19:11:43.291538692 +0000 UTC m=+1447.324150224" observedRunningTime="2026-03-17 19:11:44.23791315 +0000 UTC m=+1448.270524682" watchObservedRunningTime="2026-03-17 19:11:44.243456651 +0000 UTC m=+1448.276068173" Mar 17 19:11:44 crc kubenswrapper[5110]: I0317 19:11:44.297865 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s"] Mar 17 19:11:44 crc kubenswrapper[5110]: W0317 19:11:44.301457 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c8d3727_791f_4128_ad7b_05105eac0743.slice/crio-71b3cf7f791e6671c483e5cc8271f03027fdd315940c94dd09ab02bbc602bb38 WatchSource:0}: Error finding container 71b3cf7f791e6671c483e5cc8271f03027fdd315940c94dd09ab02bbc602bb38: Status 404 returned error can't find the container with id 71b3cf7f791e6671c483e5cc8271f03027fdd315940c94dd09ab02bbc602bb38 Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.059992 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq"] Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.071721 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.074523 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-event-sg-core-configmap\"" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.079348 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq"] Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.193945 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerStarted","Data":"8d3780f2832aea2fa4e4d3a76194556f5a1f80ba9133e57b7bcc767d46eb7aa4"} Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.193984 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerStarted","Data":"a0acdbb0fd6f3914dbdf214adde09a7b73e154a44d4b35abc69b184a048299a1"} Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.193995 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerStarted","Data":"71b3cf7f791e6671c483e5cc8271f03027fdd315940c94dd09ab02bbc602bb38"} Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.195850 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/452d004c-e213-4bb1-a044-57ba307083a9-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.195912 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/452d004c-e213-4bb1-a044-57ba307083a9-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.196002 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztq6d\" (UniqueName: \"kubernetes.io/projected/452d004c-e213-4bb1-a044-57ba307083a9-kube-api-access-ztq6d\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.196082 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/452d004c-e213-4bb1-a044-57ba307083a9-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.218767 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" podStartSLOduration=1.9212787119999999 podStartE2EDuration="2.218746818s" podCreationTimestamp="2026-03-17 19:11:43 +0000 UTC" firstStartedPulling="2026-03-17 19:11:44.303046115 +0000 UTC m=+1448.335657637" lastFinishedPulling="2026-03-17 19:11:44.600514221 +0000 UTC m=+1448.633125743" observedRunningTime="2026-03-17 19:11:45.211940842 +0000 UTC m=+1449.244552384" watchObservedRunningTime="2026-03-17 19:11:45.218746818 +0000 UTC m=+1449.251358340" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.297146 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/452d004c-e213-4bb1-a044-57ba307083a9-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.297262 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/452d004c-e213-4bb1-a044-57ba307083a9-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.297315 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/452d004c-e213-4bb1-a044-57ba307083a9-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.297450 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ztq6d\" (UniqueName: \"kubernetes.io/projected/452d004c-e213-4bb1-a044-57ba307083a9-kube-api-access-ztq6d\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.298719 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/452d004c-e213-4bb1-a044-57ba307083a9-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.298785 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/452d004c-e213-4bb1-a044-57ba307083a9-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.306165 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/452d004c-e213-4bb1-a044-57ba307083a9-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.322809 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztq6d\" (UniqueName: \"kubernetes.io/projected/452d004c-e213-4bb1-a044-57ba307083a9-kube-api-access-ztq6d\") pod \"default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq\" (UID: \"452d004c-e213-4bb1-a044-57ba307083a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.388514 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" Mar 17 19:11:45 crc kubenswrapper[5110]: I0317 19:11:45.591348 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq"] Mar 17 19:11:45 crc kubenswrapper[5110]: W0317 19:11:45.600504 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod452d004c_e213_4bb1_a044_57ba307083a9.slice/crio-4c7d90f11fb5997842c79cc45fb316d8a15a8590e79a2a7968f3e61f78a320a6 WatchSource:0}: Error finding container 4c7d90f11fb5997842c79cc45fb316d8a15a8590e79a2a7968f3e61f78a320a6: Status 404 returned error can't find the container with id 4c7d90f11fb5997842c79cc45fb316d8a15a8590e79a2a7968f3e61f78a320a6 Mar 17 19:11:46 crc kubenswrapper[5110]: I0317 19:11:46.203456 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerStarted","Data":"2f6a725219092775506bf8a31e72fadb497aa4889a4cff452a5363457f0c8316"} Mar 17 19:11:46 crc kubenswrapper[5110]: I0317 19:11:46.204902 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerStarted","Data":"ea84265e7af072b5dd11218b1a87811c169c7af92c7eedbc9e2e204a20e4ccd9"} Mar 17 19:11:46 crc kubenswrapper[5110]: I0317 19:11:46.205010 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerStarted","Data":"4c7d90f11fb5997842c79cc45fb316d8a15a8590e79a2a7968f3e61f78a320a6"} Mar 17 19:11:49 crc kubenswrapper[5110]: I0317 19:11:49.488044 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:49 crc kubenswrapper[5110]: I0317 19:11:49.522284 5110 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:49 crc kubenswrapper[5110]: I0317 19:11:49.554511 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" podStartSLOduration=4.1759501199999995 podStartE2EDuration="4.554492456s" podCreationTimestamp="2026-03-17 19:11:45 +0000 UTC" firstStartedPulling="2026-03-17 19:11:45.602040542 +0000 UTC m=+1449.634652064" lastFinishedPulling="2026-03-17 19:11:45.980582878 +0000 UTC m=+1450.013194400" observedRunningTime="2026-03-17 19:11:47.235027661 +0000 UTC m=+1451.267639193" watchObservedRunningTime="2026-03-17 19:11:49.554492456 +0000 UTC m=+1453.587103978" Mar 17 19:11:50 crc kubenswrapper[5110]: I0317 19:11:50.278327 5110 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Mar 17 19:11:56 crc kubenswrapper[5110]: I0317 19:11:56.678837 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-qm4tl"] Mar 17 19:11:56 crc kubenswrapper[5110]: I0317 19:11:56.679728 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" podUID="b47bd39b-545a-4b18-9a09-e85804e73449" containerName="default-interconnect" containerID="cri-o://19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d" gracePeriod=30 Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.047524 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.085661 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-f5sdx"] Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.086551 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b47bd39b-545a-4b18-9a09-e85804e73449" containerName="default-interconnect" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.086574 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="b47bd39b-545a-4b18-9a09-e85804e73449" containerName="default-interconnect" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.086796 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="b47bd39b-545a-4b18-9a09-e85804e73449" containerName="default-interconnect" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.093394 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.098081 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-f5sdx"] Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.167965 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-config\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168007 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-ca\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168047 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-ca\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168124 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-credentials\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168170 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-credentials\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168293 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6hzw\" (UniqueName: \"kubernetes.io/projected/b47bd39b-545a-4b18-9a09-e85804e73449-kube-api-access-b6hzw\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168323 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-users\") pod \"b47bd39b-545a-4b18-9a09-e85804e73449\" (UID: \"b47bd39b-545a-4b18-9a09-e85804e73449\") " Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168450 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmhq\" (UniqueName: \"kubernetes.io/projected/a82b157f-c3cf-434c-876c-630af0ed199b-kube-api-access-cvmhq\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168519 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168549 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/a82b157f-c3cf-434c-876c-630af0ed199b-sasl-config\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168621 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168657 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-sasl-users\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168676 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168692 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.168854 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.173939 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.174025 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.174079 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.174725 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.175187 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.186370 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b47bd39b-545a-4b18-9a09-e85804e73449-kube-api-access-b6hzw" (OuterVolumeSpecName: "kube-api-access-b6hzw") pod "b47bd39b-545a-4b18-9a09-e85804e73449" (UID: "b47bd39b-545a-4b18-9a09-e85804e73449"). InnerVolumeSpecName "kube-api-access-b6hzw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.269830 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmhq\" (UniqueName: \"kubernetes.io/projected/a82b157f-c3cf-434c-876c-630af0ed199b-kube-api-access-cvmhq\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.269905 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.269941 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/a82b157f-c3cf-434c-876c-630af0ed199b-sasl-config\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.269996 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270040 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-sasl-users\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270080 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270100 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270157 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-b6hzw\" (UniqueName: \"kubernetes.io/projected/b47bd39b-545a-4b18-9a09-e85804e73449-kube-api-access-b6hzw\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270172 5110 reconciler_common.go:299] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-users\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270184 5110 reconciler_common.go:299] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b47bd39b-545a-4b18-9a09-e85804e73449-sasl-config\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270196 5110 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270209 5110 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270222 5110 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.270236 5110 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b47bd39b-545a-4b18-9a09-e85804e73449-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.271587 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/a82b157f-c3cf-434c-876c-630af0ed199b-sasl-config\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.273755 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-sasl-users\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.274429 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.274679 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.278809 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.279458 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/a82b157f-c3cf-434c-876c-630af0ed199b-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.283956 5110 generic.go:358] "Generic (PLEG): container finished" podID="b47bd39b-545a-4b18-9a09-e85804e73449" containerID="19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d" exitCode=0 Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.284075 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.284113 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" event={"ID":"b47bd39b-545a-4b18-9a09-e85804e73449","Type":"ContainerDied","Data":"19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d"} Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.284144 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-qm4tl" event={"ID":"b47bd39b-545a-4b18-9a09-e85804e73449","Type":"ContainerDied","Data":"625f6492479425830135ed971fd7e95fc96b1a1a11199141ea843156ad0992f8"} Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.284163 5110 scope.go:117] "RemoveContainer" containerID="19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.286754 5110 generic.go:358] "Generic (PLEG): container finished" podID="091754de-5b5e-438e-ab68-395e08cdccb3" containerID="31cd2b73b7ad916f59595243040588ed6a5cf305b4e38b4ccad1a81faee2bc8c" exitCode=0 Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.286877 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerDied","Data":"31cd2b73b7ad916f59595243040588ed6a5cf305b4e38b4ccad1a81faee2bc8c"} Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.287484 5110 scope.go:117] "RemoveContainer" containerID="31cd2b73b7ad916f59595243040588ed6a5cf305b4e38b4ccad1a81faee2bc8c" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.294904 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmhq\" (UniqueName: \"kubernetes.io/projected/a82b157f-c3cf-434c-876c-630af0ed199b-kube-api-access-cvmhq\") pod \"default-interconnect-55bf8d5cb-f5sdx\" (UID: \"a82b157f-c3cf-434c-876c-630af0ed199b\") " pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.296080 5110 generic.go:358] "Generic (PLEG): container finished" podID="452d004c-e213-4bb1-a044-57ba307083a9" containerID="ea84265e7af072b5dd11218b1a87811c169c7af92c7eedbc9e2e204a20e4ccd9" exitCode=0 Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.296151 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerDied","Data":"ea84265e7af072b5dd11218b1a87811c169c7af92c7eedbc9e2e204a20e4ccd9"} Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.296592 5110 scope.go:117] "RemoveContainer" containerID="ea84265e7af072b5dd11218b1a87811c169c7af92c7eedbc9e2e204a20e4ccd9" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.317223 5110 scope.go:117] "RemoveContainer" containerID="19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d" Mar 17 19:11:57 crc kubenswrapper[5110]: E0317 19:11:57.318866 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d\": container with ID starting with 19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d not found: ID does not exist" containerID="19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.318959 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d"} err="failed to get container status \"19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d\": rpc error: code = NotFound desc = could not find container \"19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d\": container with ID starting with 19a1b54e95ad75ea72e7e9caff232e62459b1006e190ebb59d4805bf6bb2f82d not found: ID does not exist" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.374252 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-qm4tl"] Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.385753 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-qm4tl"] Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.412412 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" Mar 17 19:11:57 crc kubenswrapper[5110]: I0317 19:11:57.662539 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-f5sdx"] Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.304395 5110 generic.go:358] "Generic (PLEG): container finished" podID="ddba8057-d741-42e2-8e66-d1a5f0ed2524" containerID="1e3df99559593b3cbeb74186607aaa2bce06c867a2454c2252cb36b19734502a" exitCode=0 Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.304492 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerDied","Data":"1e3df99559593b3cbeb74186607aaa2bce06c867a2454c2252cb36b19734502a"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.305255 5110 scope.go:117] "RemoveContainer" containerID="1e3df99559593b3cbeb74186607aaa2bce06c867a2454c2252cb36b19734502a" Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.309621 5110 generic.go:358] "Generic (PLEG): container finished" podID="e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad" containerID="77147d07f1f29426cd72ef9ab476deb9f2f2bc446b35adadc496222fa940ae56" exitCode=0 Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.309688 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerDied","Data":"77147d07f1f29426cd72ef9ab476deb9f2f2bc446b35adadc496222fa940ae56"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.310202 5110 scope.go:117] "RemoveContainer" containerID="77147d07f1f29426cd72ef9ab476deb9f2f2bc446b35adadc496222fa940ae56" Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.315222 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerStarted","Data":"53ee0a1e1f0bc4b6a50831da2c138111f639d031df54c7c48b8f25e80e100903"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.321283 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" event={"ID":"a82b157f-c3cf-434c-876c-630af0ed199b","Type":"ContainerStarted","Data":"af0e3c4c68f4152e6ae140e1cb38f32f9d373ba639d2b8a8c8e80a6a7084328c"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.321311 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" event={"ID":"a82b157f-c3cf-434c-876c-630af0ed199b","Type":"ContainerStarted","Data":"c2b7560d79d13bd06608d5516fc9651f1965b4e49ff272468b08c31680380349"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.333092 5110 generic.go:358] "Generic (PLEG): container finished" podID="9c8d3727-791f-4128-ad7b-05105eac0743" containerID="a0acdbb0fd6f3914dbdf214adde09a7b73e154a44d4b35abc69b184a048299a1" exitCode=0 Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.333226 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerDied","Data":"a0acdbb0fd6f3914dbdf214adde09a7b73e154a44d4b35abc69b184a048299a1"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.333675 5110 scope.go:117] "RemoveContainer" containerID="a0acdbb0fd6f3914dbdf214adde09a7b73e154a44d4b35abc69b184a048299a1" Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.342794 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerStarted","Data":"99274e7119050b26143ae1fa29da8c16e0da189fd390836f8a30c94af2d366a3"} Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.366376 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-55bf8d5cb-f5sdx" podStartSLOduration=2.366351197 podStartE2EDuration="2.366351197s" podCreationTimestamp="2026-03-17 19:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:11:58.362312837 +0000 UTC m=+1462.394924389" watchObservedRunningTime="2026-03-17 19:11:58.366351197 +0000 UTC m=+1462.398962719" Mar 17 19:11:58 crc kubenswrapper[5110]: I0317 19:11:58.753716 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b47bd39b-545a-4b18-9a09-e85804e73449" path="/var/lib/kubelet/pods/b47bd39b-545a-4b18-9a09-e85804e73449/volumes" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.355033 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerStarted","Data":"01dad42353ec6fc031f4d2434f91a8dff8cae33cd91936c58bbc09fc92366c4f"} Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.361083 5110 generic.go:358] "Generic (PLEG): container finished" podID="091754de-5b5e-438e-ab68-395e08cdccb3" containerID="99274e7119050b26143ae1fa29da8c16e0da189fd390836f8a30c94af2d366a3" exitCode=0 Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.361173 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerDied","Data":"99274e7119050b26143ae1fa29da8c16e0da189fd390836f8a30c94af2d366a3"} Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.361457 5110 scope.go:117] "RemoveContainer" containerID="31cd2b73b7ad916f59595243040588ed6a5cf305b4e38b4ccad1a81faee2bc8c" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.362192 5110 scope.go:117] "RemoveContainer" containerID="99274e7119050b26143ae1fa29da8c16e0da189fd390836f8a30c94af2d366a3" Mar 17 19:11:59 crc kubenswrapper[5110]: E0317 19:11:59.362965 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg_service-telemetry(091754de-5b5e-438e-ab68-395e08cdccb3)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" podUID="091754de-5b5e-438e-ab68-395e08cdccb3" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.383457 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerStarted","Data":"4286c9c30ad5f72d2fd64238540c6b68aee1fe89221930b051e28e4e7eede1c2"} Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.414324 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerStarted","Data":"f5ce98eadade6d10839f8fafcaa2f8c1178ec38392f62c0a327ecbad15709004"} Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.432557 5110 generic.go:358] "Generic (PLEG): container finished" podID="452d004c-e213-4bb1-a044-57ba307083a9" containerID="53ee0a1e1f0bc4b6a50831da2c138111f639d031df54c7c48b8f25e80e100903" exitCode=0 Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.433614 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerDied","Data":"53ee0a1e1f0bc4b6a50831da2c138111f639d031df54c7c48b8f25e80e100903"} Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.433659 5110 scope.go:117] "RemoveContainer" containerID="ea84265e7af072b5dd11218b1a87811c169c7af92c7eedbc9e2e204a20e4ccd9" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.433954 5110 scope.go:117] "RemoveContainer" containerID="53ee0a1e1f0bc4b6a50831da2c138111f639d031df54c7c48b8f25e80e100903" Mar 17 19:11:59 crc kubenswrapper[5110]: E0317 19:11:59.434178 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq_service-telemetry(452d004c-e213-4bb1-a044-57ba307083a9)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" podUID="452d004c-e213-4bb1-a044-57ba307083a9" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.609964 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.616424 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.621020 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.621992 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-selfsigned\"" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.622014 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"qdr-test-config\"" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.723197 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/76b9f4d4-d227-46b2-b023-9b1956d76179-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.723355 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26q5q\" (UniqueName: \"kubernetes.io/projected/76b9f4d4-d227-46b2-b023-9b1956d76179-kube-api-access-26q5q\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.723422 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/76b9f4d4-d227-46b2-b023-9b1956d76179-qdr-test-config\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.825324 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-26q5q\" (UniqueName: \"kubernetes.io/projected/76b9f4d4-d227-46b2-b023-9b1956d76179-kube-api-access-26q5q\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.825567 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/76b9f4d4-d227-46b2-b023-9b1956d76179-qdr-test-config\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.825912 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/76b9f4d4-d227-46b2-b023-9b1956d76179-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.826822 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/76b9f4d4-d227-46b2-b023-9b1956d76179-qdr-test-config\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.842744 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/76b9f4d4-d227-46b2-b023-9b1956d76179-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.848909 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-26q5q\" (UniqueName: \"kubernetes.io/projected/76b9f4d4-d227-46b2-b023-9b1956d76179-kube-api-access-26q5q\") pod \"qdr-test\" (UID: \"76b9f4d4-d227-46b2-b023-9b1956d76179\") " pod="service-telemetry/qdr-test" Mar 17 19:11:59 crc kubenswrapper[5110]: I0317 19:11:59.951128 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.126132 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562912-mvgwc"] Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.138112 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562912-mvgwc"] Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.138234 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.140637 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.141589 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.143275 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.180045 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.231784 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfhjx\" (UniqueName: \"kubernetes.io/projected/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4-kube-api-access-dfhjx\") pod \"auto-csr-approver-29562912-mvgwc\" (UID: \"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4\") " pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.333784 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dfhjx\" (UniqueName: \"kubernetes.io/projected/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4-kube-api-access-dfhjx\") pod \"auto-csr-approver-29562912-mvgwc\" (UID: \"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4\") " pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.353547 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfhjx\" (UniqueName: \"kubernetes.io/projected/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4-kube-api-access-dfhjx\") pod \"auto-csr-approver-29562912-mvgwc\" (UID: \"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4\") " pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.442155 5110 generic.go:358] "Generic (PLEG): container finished" podID="9c8d3727-791f-4128-ad7b-05105eac0743" containerID="01dad42353ec6fc031f4d2434f91a8dff8cae33cd91936c58bbc09fc92366c4f" exitCode=0 Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.442237 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerDied","Data":"01dad42353ec6fc031f4d2434f91a8dff8cae33cd91936c58bbc09fc92366c4f"} Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.442319 5110 scope.go:117] "RemoveContainer" containerID="a0acdbb0fd6f3914dbdf214adde09a7b73e154a44d4b35abc69b184a048299a1" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.443027 5110 scope.go:117] "RemoveContainer" containerID="01dad42353ec6fc031f4d2434f91a8dff8cae33cd91936c58bbc09fc92366c4f" Mar 17 19:12:00 crc kubenswrapper[5110]: E0317 19:12:00.443440 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-59876f8548-vkd7s_service-telemetry(9c8d3727-791f-4128-ad7b-05105eac0743)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" podUID="9c8d3727-791f-4128-ad7b-05105eac0743" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.450695 5110 generic.go:358] "Generic (PLEG): container finished" podID="ddba8057-d741-42e2-8e66-d1a5f0ed2524" containerID="4286c9c30ad5f72d2fd64238540c6b68aee1fe89221930b051e28e4e7eede1c2" exitCode=0 Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.450792 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerDied","Data":"4286c9c30ad5f72d2fd64238540c6b68aee1fe89221930b051e28e4e7eede1c2"} Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.451911 5110 scope.go:117] "RemoveContainer" containerID="4286c9c30ad5f72d2fd64238540c6b68aee1fe89221930b051e28e4e7eede1c2" Mar 17 19:12:00 crc kubenswrapper[5110]: E0317 19:12:00.452278 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf_service-telemetry(ddba8057-d741-42e2-8e66-d1a5f0ed2524)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" podUID="ddba8057-d741-42e2-8e66-d1a5f0ed2524" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.458447 5110 generic.go:358] "Generic (PLEG): container finished" podID="e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad" containerID="f5ce98eadade6d10839f8fafcaa2f8c1178ec38392f62c0a327ecbad15709004" exitCode=0 Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.458530 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerDied","Data":"f5ce98eadade6d10839f8fafcaa2f8c1178ec38392f62c0a327ecbad15709004"} Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.459219 5110 scope.go:117] "RemoveContainer" containerID="f5ce98eadade6d10839f8fafcaa2f8c1178ec38392f62c0a327ecbad15709004" Mar 17 19:12:00 crc kubenswrapper[5110]: E0317 19:12:00.459549 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh_service-telemetry(e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" podUID="e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.465114 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"76b9f4d4-d227-46b2-b023-9b1956d76179","Type":"ContainerStarted","Data":"1cb35536e5618bc6e0c1b70ac0a1fb6a12bdb978c9720b27a020ff9ec640a2ed"} Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.467997 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.532503 5110 scope.go:117] "RemoveContainer" containerID="1e3df99559593b3cbeb74186607aaa2bce06c867a2454c2252cb36b19734502a" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.618283 5110 scope.go:117] "RemoveContainer" containerID="77147d07f1f29426cd72ef9ab476deb9f2f2bc446b35adadc496222fa940ae56" Mar 17 19:12:00 crc kubenswrapper[5110]: I0317 19:12:00.970636 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562912-mvgwc"] Mar 17 19:12:00 crc kubenswrapper[5110]: W0317 19:12:00.983015 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb686fe0d_dc7d_4a76_82c4_2ee230ead1f4.slice/crio-681fe378d1d28a659fc97ecbcb47c0e26929c690179703c01ef076c0ec86c55b WatchSource:0}: Error finding container 681fe378d1d28a659fc97ecbcb47c0e26929c690179703c01ef076c0ec86c55b: Status 404 returned error can't find the container with id 681fe378d1d28a659fc97ecbcb47c0e26929c690179703c01ef076c0ec86c55b Mar 17 19:12:01 crc kubenswrapper[5110]: I0317 19:12:01.474969 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" event={"ID":"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4","Type":"ContainerStarted","Data":"681fe378d1d28a659fc97ecbcb47c0e26929c690179703c01ef076c0ec86c55b"} Mar 17 19:12:03 crc kubenswrapper[5110]: I0317 19:12:03.498880 5110 generic.go:358] "Generic (PLEG): container finished" podID="b686fe0d-dc7d-4a76-82c4-2ee230ead1f4" containerID="2f3f75187608449518626191a2d3bd86686046af490651692ca73b2d14207936" exitCode=0 Mar 17 19:12:03 crc kubenswrapper[5110]: I0317 19:12:03.499088 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" event={"ID":"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4","Type":"ContainerDied","Data":"2f3f75187608449518626191a2d3bd86686046af490651692ca73b2d14207936"} Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.141898 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.263583 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfhjx\" (UniqueName: \"kubernetes.io/projected/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4-kube-api-access-dfhjx\") pod \"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4\" (UID: \"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4\") " Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.274261 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4-kube-api-access-dfhjx" (OuterVolumeSpecName: "kube-api-access-dfhjx") pod "b686fe0d-dc7d-4a76-82c4-2ee230ead1f4" (UID: "b686fe0d-dc7d-4a76-82c4-2ee230ead1f4"). InnerVolumeSpecName "kube-api-access-dfhjx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.365359 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dfhjx\" (UniqueName: \"kubernetes.io/projected/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4-kube-api-access-dfhjx\") on node \"crc\" DevicePath \"\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.540544 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"76b9f4d4-d227-46b2-b023-9b1956d76179","Type":"ContainerStarted","Data":"747bd98124a454026de79a4fdcdf6bceb5908dd61334ff8796d0a4308d7e7d2d"} Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.542522 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" event={"ID":"b686fe0d-dc7d-4a76-82c4-2ee230ead1f4","Type":"ContainerDied","Data":"681fe378d1d28a659fc97ecbcb47c0e26929c690179703c01ef076c0ec86c55b"} Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.542556 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="681fe378d1d28a659fc97ecbcb47c0e26929c690179703c01ef076c0ec86c55b" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.542562 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562912-mvgwc" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.558398 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=1.590538217 podStartE2EDuration="10.558380818s" podCreationTimestamp="2026-03-17 19:11:59 +0000 UTC" firstStartedPulling="2026-03-17 19:12:00.183248268 +0000 UTC m=+1464.217242638" lastFinishedPulling="2026-03-17 19:12:09.152473717 +0000 UTC m=+1473.185085239" observedRunningTime="2026-03-17 19:12:09.554833092 +0000 UTC m=+1473.587444614" watchObservedRunningTime="2026-03-17 19:12:09.558380818 +0000 UTC m=+1473.590992340" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.822427 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-m2r9s"] Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.823335 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b686fe0d-dc7d-4a76-82c4-2ee230ead1f4" containerName="oc" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.823361 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="b686fe0d-dc7d-4a76-82c4-2ee230ead1f4" containerName="oc" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.823527 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="b686fe0d-dc7d-4a76-82c4-2ee230ead1f4" containerName="oc" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.830479 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.832369 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-collectd-config\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.832479 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-sensubility-config\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.832950 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-healthcheck-log\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.833231 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-collectd-entrypoint-script\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.833515 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-ceilometer-publisher\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.833866 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-ceilometer-entrypoint-script\"" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.836477 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-m2r9s"] Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973600 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-sensubility-config\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973678 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-config\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973712 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-healthcheck-log\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973747 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-publisher\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973783 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973824 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:09 crc kubenswrapper[5110]: I0317 19:12:09.973916 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvzxz\" (UniqueName: \"kubernetes.io/projected/58dbad41-8a71-475a-91cb-c18dd788dceb-kube-api-access-vvzxz\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.074805 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vvzxz\" (UniqueName: \"kubernetes.io/projected/58dbad41-8a71-475a-91cb-c18dd788dceb-kube-api-access-vvzxz\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.074895 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-sensubility-config\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.074919 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-config\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.074937 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-healthcheck-log\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.075136 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-publisher\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.075249 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.075352 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.075882 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-healthcheck-log\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.075943 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-config\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.076046 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-publisher\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.076364 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.076887 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.077131 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-sensubility-config\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.092426 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvzxz\" (UniqueName: \"kubernetes.io/projected/58dbad41-8a71-475a-91cb-c18dd788dceb-kube-api-access-vvzxz\") pod \"stf-smoketest-smoke1-m2r9s\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.143847 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.202859 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.217855 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.217991 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.224492 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562906-vm2mw"] Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.230628 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562906-vm2mw"] Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.383492 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw9qs\" (UniqueName: \"kubernetes.io/projected/1a207b29-155d-4d14-ba09-f2af999798e4-kube-api-access-kw9qs\") pod \"curl\" (UID: \"1a207b29-155d-4d14-ba09-f2af999798e4\") " pod="service-telemetry/curl" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.417572 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-m2r9s"] Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.485252 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kw9qs\" (UniqueName: \"kubernetes.io/projected/1a207b29-155d-4d14-ba09-f2af999798e4-kube-api-access-kw9qs\") pod \"curl\" (UID: \"1a207b29-155d-4d14-ba09-f2af999798e4\") " pod="service-telemetry/curl" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.510257 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw9qs\" (UniqueName: \"kubernetes.io/projected/1a207b29-155d-4d14-ba09-f2af999798e4-kube-api-access-kw9qs\") pod \"curl\" (UID: \"1a207b29-155d-4d14-ba09-f2af999798e4\") " pod="service-telemetry/curl" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.551049 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" event={"ID":"58dbad41-8a71-475a-91cb-c18dd788dceb","Type":"ContainerStarted","Data":"da9d9ecd7ab3819792376e952c9885be6f7d60afffe0752d91f543084ad4e373"} Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.575689 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.754810 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0e3708-d07f-4103-b22b-ba3fa96d0aed" path="/var/lib/kubelet/pods/7c0e3708-d07f-4103-b22b-ba3fa96d0aed/volumes" Mar 17 19:12:10 crc kubenswrapper[5110]: I0317 19:12:10.792363 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 17 19:12:11 crc kubenswrapper[5110]: I0317 19:12:11.562426 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"1a207b29-155d-4d14-ba09-f2af999798e4","Type":"ContainerStarted","Data":"e629ec443ba6fb264c32a1137a3fb33b80fa2af3497600418ae2646f7b7cdf46"} Mar 17 19:12:11 crc kubenswrapper[5110]: I0317 19:12:11.744312 5110 scope.go:117] "RemoveContainer" containerID="01dad42353ec6fc031f4d2434f91a8dff8cae33cd91936c58bbc09fc92366c4f" Mar 17 19:12:13 crc kubenswrapper[5110]: I0317 19:12:13.588545 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-59876f8548-vkd7s" event={"ID":"9c8d3727-791f-4128-ad7b-05105eac0743","Type":"ContainerStarted","Data":"cfccc8d2d6f10cbfd5d0d4a5ecf16bd43589515058fd846605c65e3d1b0284b4"} Mar 17 19:12:13 crc kubenswrapper[5110]: I0317 19:12:13.590042 5110 generic.go:358] "Generic (PLEG): container finished" podID="1a207b29-155d-4d14-ba09-f2af999798e4" containerID="9f4dd0a52da1e14253243bf9b7b0fd3540dd5a36fabc9fba4a8228ffb5ead7e7" exitCode=0 Mar 17 19:12:13 crc kubenswrapper[5110]: I0317 19:12:13.590192 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"1a207b29-155d-4d14-ba09-f2af999798e4","Type":"ContainerDied","Data":"9f4dd0a52da1e14253243bf9b7b0fd3540dd5a36fabc9fba4a8228ffb5ead7e7"} Mar 17 19:12:13 crc kubenswrapper[5110]: I0317 19:12:13.744671 5110 scope.go:117] "RemoveContainer" containerID="99274e7119050b26143ae1fa29da8c16e0da189fd390836f8a30c94af2d366a3" Mar 17 19:12:13 crc kubenswrapper[5110]: I0317 19:12:13.745154 5110 scope.go:117] "RemoveContainer" containerID="4286c9c30ad5f72d2fd64238540c6b68aee1fe89221930b051e28e4e7eede1c2" Mar 17 19:12:14 crc kubenswrapper[5110]: I0317 19:12:14.746455 5110 scope.go:117] "RemoveContainer" containerID="53ee0a1e1f0bc4b6a50831da2c138111f639d031df54c7c48b8f25e80e100903" Mar 17 19:12:14 crc kubenswrapper[5110]: I0317 19:12:14.746817 5110 scope.go:117] "RemoveContainer" containerID="f5ce98eadade6d10839f8fafcaa2f8c1178ec38392f62c0a327ecbad15709004" Mar 17 19:12:16 crc kubenswrapper[5110]: I0317 19:12:16.686323 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 17 19:12:16 crc kubenswrapper[5110]: I0317 19:12:16.776917 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw9qs\" (UniqueName: \"kubernetes.io/projected/1a207b29-155d-4d14-ba09-f2af999798e4-kube-api-access-kw9qs\") pod \"1a207b29-155d-4d14-ba09-f2af999798e4\" (UID: \"1a207b29-155d-4d14-ba09-f2af999798e4\") " Mar 17 19:12:16 crc kubenswrapper[5110]: I0317 19:12:16.784456 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a207b29-155d-4d14-ba09-f2af999798e4-kube-api-access-kw9qs" (OuterVolumeSpecName: "kube-api-access-kw9qs") pod "1a207b29-155d-4d14-ba09-f2af999798e4" (UID: "1a207b29-155d-4d14-ba09-f2af999798e4"). InnerVolumeSpecName "kube-api-access-kw9qs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:12:16 crc kubenswrapper[5110]: I0317 19:12:16.844999 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_1a207b29-155d-4d14-ba09-f2af999798e4/curl/0.log" Mar 17 19:12:16 crc kubenswrapper[5110]: I0317 19:12:16.879124 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kw9qs\" (UniqueName: \"kubernetes.io/projected/1a207b29-155d-4d14-ba09-f2af999798e4-kube-api-access-kw9qs\") on node \"crc\" DevicePath \"\"" Mar 17 19:12:17 crc kubenswrapper[5110]: I0317 19:12:17.080262 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-694dc457d5-rcxnv_4a47a709-2c00-40b6-bd44-90ee935a14f5/prometheus-webhook-snmp/0.log" Mar 17 19:12:17 crc kubenswrapper[5110]: I0317 19:12:17.627194 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"1a207b29-155d-4d14-ba09-f2af999798e4","Type":"ContainerDied","Data":"e629ec443ba6fb264c32a1137a3fb33b80fa2af3497600418ae2646f7b7cdf46"} Mar 17 19:12:17 crc kubenswrapper[5110]: I0317 19:12:17.627239 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e629ec443ba6fb264c32a1137a3fb33b80fa2af3497600418ae2646f7b7cdf46" Mar 17 19:12:17 crc kubenswrapper[5110]: I0317 19:12:17.627314 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 17 19:12:21 crc kubenswrapper[5110]: I0317 19:12:21.664288 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg" event={"ID":"091754de-5b5e-438e-ab68-395e08cdccb3","Type":"ContainerStarted","Data":"883482689f12d0fb26f603adc3cb792255cd45f2f48f4c72daea5e8d04e80648"} Mar 17 19:12:21 crc kubenswrapper[5110]: I0317 19:12:21.667094 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf" event={"ID":"ddba8057-d741-42e2-8e66-d1a5f0ed2524","Type":"ContainerStarted","Data":"cc3d93998cf380e2cfe4043749914598f8df10180f3791635f3acff8a99503fb"} Mar 17 19:12:21 crc kubenswrapper[5110]: I0317 19:12:21.678780 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh" event={"ID":"e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad","Type":"ContainerStarted","Data":"3d5f6b3c252e6c03d5fe11eabe76506f6aaa2cf4b772fbd74800178c3138465d"} Mar 17 19:12:21 crc kubenswrapper[5110]: I0317 19:12:21.682568 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq" event={"ID":"452d004c-e213-4bb1-a044-57ba307083a9","Type":"ContainerStarted","Data":"d9918758d04a6884aeff1334054ee334b2112fc6056c3c0b7cdff15efeb89036"} Mar 17 19:12:21 crc kubenswrapper[5110]: I0317 19:12:21.685374 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" event={"ID":"58dbad41-8a71-475a-91cb-c18dd788dceb","Type":"ContainerStarted","Data":"b4e881f1f67d7e0cbafddd5290829b38e5d5d2989bbcf74e06615cb7f992547e"} Mar 17 19:12:28 crc kubenswrapper[5110]: I0317 19:12:28.753763 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" event={"ID":"58dbad41-8a71-475a-91cb-c18dd788dceb","Type":"ContainerStarted","Data":"5d13909bf0427a9d817e201f30ff39fa59cdda107bd654d5842cfe75eafb5d37"} Mar 17 19:12:28 crc kubenswrapper[5110]: I0317 19:12:28.776242 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" podStartSLOduration=2.04429099 podStartE2EDuration="19.776223311s" podCreationTimestamp="2026-03-17 19:12:09 +0000 UTC" firstStartedPulling="2026-03-17 19:12:10.4233962 +0000 UTC m=+1474.456007712" lastFinishedPulling="2026-03-17 19:12:28.155328511 +0000 UTC m=+1492.187940033" observedRunningTime="2026-03-17 19:12:28.771486191 +0000 UTC m=+1492.804097713" watchObservedRunningTime="2026-03-17 19:12:28.776223311 +0000 UTC m=+1492.808834823" Mar 17 19:12:37 crc kubenswrapper[5110]: I0317 19:12:37.522003 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:12:37 crc kubenswrapper[5110]: I0317 19:12:37.531645 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:12:37 crc kubenswrapper[5110]: I0317 19:12:37.539230 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:12:37 crc kubenswrapper[5110]: I0317 19:12:37.544172 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:12:42 crc kubenswrapper[5110]: I0317 19:12:42.133466 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:12:42 crc kubenswrapper[5110]: I0317 19:12:42.134098 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:12:47 crc kubenswrapper[5110]: I0317 19:12:47.254976 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-694dc457d5-rcxnv_4a47a709-2c00-40b6-bd44-90ee935a14f5/prometheus-webhook-snmp/0.log" Mar 17 19:12:48 crc kubenswrapper[5110]: I0317 19:12:48.848295 5110 scope.go:117] "RemoveContainer" containerID="35d758205c4e3e118fd55027d0804dff72a46218879b69c30d269fbc907d5524" Mar 17 19:12:55 crc kubenswrapper[5110]: I0317 19:12:55.981722 5110 generic.go:358] "Generic (PLEG): container finished" podID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerID="b4e881f1f67d7e0cbafddd5290829b38e5d5d2989bbcf74e06615cb7f992547e" exitCode=0 Mar 17 19:12:55 crc kubenswrapper[5110]: I0317 19:12:55.981906 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" event={"ID":"58dbad41-8a71-475a-91cb-c18dd788dceb","Type":"ContainerDied","Data":"b4e881f1f67d7e0cbafddd5290829b38e5d5d2989bbcf74e06615cb7f992547e"} Mar 17 19:12:55 crc kubenswrapper[5110]: I0317 19:12:55.983574 5110 scope.go:117] "RemoveContainer" containerID="b4e881f1f67d7e0cbafddd5290829b38e5d5d2989bbcf74e06615cb7f992547e" Mar 17 19:13:01 crc kubenswrapper[5110]: I0317 19:13:01.022460 5110 generic.go:358] "Generic (PLEG): container finished" podID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerID="5d13909bf0427a9d817e201f30ff39fa59cdda107bd654d5842cfe75eafb5d37" exitCode=0 Mar 17 19:13:01 crc kubenswrapper[5110]: I0317 19:13:01.022549 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" event={"ID":"58dbad41-8a71-475a-91cb-c18dd788dceb","Type":"ContainerDied","Data":"5d13909bf0427a9d817e201f30ff39fa59cdda107bd654d5842cfe75eafb5d37"} Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.305546 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.425576 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-healthcheck-log\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.425776 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvzxz\" (UniqueName: \"kubernetes.io/projected/58dbad41-8a71-475a-91cb-c18dd788dceb-kube-api-access-vvzxz\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.425872 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-sensubility-config\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.425976 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-publisher\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.426159 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-config\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.426269 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-entrypoint-script\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.426973 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-entrypoint-script\") pod \"58dbad41-8a71-475a-91cb-c18dd788dceb\" (UID: \"58dbad41-8a71-475a-91cb-c18dd788dceb\") " Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.431951 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58dbad41-8a71-475a-91cb-c18dd788dceb-kube-api-access-vvzxz" (OuterVolumeSpecName: "kube-api-access-vvzxz") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "kube-api-access-vvzxz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.535151 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vvzxz\" (UniqueName: \"kubernetes.io/projected/58dbad41-8a71-475a-91cb-c18dd788dceb-kube-api-access-vvzxz\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.708937 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.708994 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.709084 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.709371 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.709964 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.722476 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "58dbad41-8a71-475a-91cb-c18dd788dceb" (UID: "58dbad41-8a71-475a-91cb-c18dd788dceb"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.738196 5110 reconciler_common.go:299] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-config\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.738231 5110 reconciler_common.go:299] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.738242 5110 reconciler_common.go:299] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.738252 5110 reconciler_common.go:299] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-healthcheck-log\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.738259 5110 reconciler_common.go:299] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-sensubility-config\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:02 crc kubenswrapper[5110]: I0317 19:13:02.738267 5110 reconciler_common.go:299] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/58dbad41-8a71-475a-91cb-c18dd788dceb-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Mar 17 19:13:03 crc kubenswrapper[5110]: I0317 19:13:03.044381 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" event={"ID":"58dbad41-8a71-475a-91cb-c18dd788dceb","Type":"ContainerDied","Data":"da9d9ecd7ab3819792376e952c9885be6f7d60afffe0752d91f543084ad4e373"} Mar 17 19:13:03 crc kubenswrapper[5110]: I0317 19:13:03.044428 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da9d9ecd7ab3819792376e952c9885be6f7d60afffe0752d91f543084ad4e373" Mar 17 19:13:03 crc kubenswrapper[5110]: I0317 19:13:03.044509 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m2r9s" Mar 17 19:13:04 crc kubenswrapper[5110]: I0317 19:13:04.311833 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-m2r9s_58dbad41-8a71-475a-91cb-c18dd788dceb/smoketest-collectd/0.log" Mar 17 19:13:04 crc kubenswrapper[5110]: I0317 19:13:04.576374 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-m2r9s_58dbad41-8a71-475a-91cb-c18dd788dceb/smoketest-ceilometer/0.log" Mar 17 19:13:04 crc kubenswrapper[5110]: I0317 19:13:04.844547 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-55bf8d5cb-f5sdx_a82b157f-c3cf-434c-876c-630af0ed199b/default-interconnect/0.log" Mar 17 19:13:05 crc kubenswrapper[5110]: I0317 19:13:05.119279 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh_e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad/bridge/2.log" Mar 17 19:13:05 crc kubenswrapper[5110]: I0317 19:13:05.390028 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7f8f5c6486-cspgh_e32a7feb-ef4c-4949-8dec-5c2d3f2f48ad/sg-core/0.log" Mar 17 19:13:05 crc kubenswrapper[5110]: I0317 19:13:05.650089 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-59876f8548-vkd7s_9c8d3727-791f-4128-ad7b-05105eac0743/bridge/2.log" Mar 17 19:13:05 crc kubenswrapper[5110]: I0317 19:13:05.913369 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-59876f8548-vkd7s_9c8d3727-791f-4128-ad7b-05105eac0743/sg-core/0.log" Mar 17 19:13:06 crc kubenswrapper[5110]: I0317 19:13:06.156379 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf_ddba8057-d741-42e2-8e66-d1a5f0ed2524/bridge/2.log" Mar 17 19:13:06 crc kubenswrapper[5110]: I0317 19:13:06.378101 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-c9f4bb7dc-fssvf_ddba8057-d741-42e2-8e66-d1a5f0ed2524/sg-core/0.log" Mar 17 19:13:06 crc kubenswrapper[5110]: I0317 19:13:06.632097 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq_452d004c-e213-4bb1-a044-57ba307083a9/bridge/2.log" Mar 17 19:13:06 crc kubenswrapper[5110]: I0317 19:13:06.912506 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-68559b6fd5-8bjxq_452d004c-e213-4bb1-a044-57ba307083a9/sg-core/0.log" Mar 17 19:13:07 crc kubenswrapper[5110]: I0317 19:13:07.190491 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg_091754de-5b5e-438e-ab68-395e08cdccb3/bridge/2.log" Mar 17 19:13:07 crc kubenswrapper[5110]: I0317 19:13:07.417647 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-58c78bbf69-9dwbg_091754de-5b5e-438e-ab68-395e08cdccb3/sg-core/0.log" Mar 17 19:13:10 crc kubenswrapper[5110]: I0317 19:13:10.846736 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-77448bb48c-q6znx_a1d461c5-fb19-46b1-8791-ead59519666e/operator/0.log" Mar 17 19:13:11 crc kubenswrapper[5110]: I0317 19:13:11.145066 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_978872ef-66de-48f8-b24d-ae6176a05c08/prometheus/0.log" Mar 17 19:13:11 crc kubenswrapper[5110]: I0317 19:13:11.415653 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_d9344573-098c-4529-a15f-b84912c0b147/elasticsearch/0.log" Mar 17 19:13:11 crc kubenswrapper[5110]: I0317 19:13:11.671342 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-694dc457d5-rcxnv_4a47a709-2c00-40b6-bd44-90ee935a14f5/prometheus-webhook-snmp/0.log" Mar 17 19:13:11 crc kubenswrapper[5110]: I0317 19:13:11.926508 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_da51ef37-5384-4ba3-87d8-7b1feadac8dc/alertmanager/0.log" Mar 17 19:13:12 crc kubenswrapper[5110]: I0317 19:13:12.133898 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:13:12 crc kubenswrapper[5110]: I0317 19:13:12.134009 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:13:25 crc kubenswrapper[5110]: I0317 19:13:25.198455 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-7d5d86f96d-jwhgs_0c2d87e6-435e-4b36-aee5-e1c92a05821d/operator/0.log" Mar 17 19:13:29 crc kubenswrapper[5110]: I0317 19:13:29.035138 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-77448bb48c-q6znx_a1d461c5-fb19-46b1-8791-ead59519666e/operator/0.log" Mar 17 19:13:29 crc kubenswrapper[5110]: I0317 19:13:29.290518 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_76b9f4d4-d227-46b2-b023-9b1956d76179/qdr/0.log" Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.132417 5110 patch_prober.go:28] interesting pod/machine-config-daemon-6jbgs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.133090 5110 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.133138 5110 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.133707 5110 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2"} pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.133761 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" containerName="machine-config-daemon" containerID="cri-o://654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" gracePeriod=600 Mar 17 19:13:42 crc kubenswrapper[5110]: E0317 19:13:42.255931 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.372312 5110 generic.go:358] "Generic (PLEG): container finished" podID="7da49224-3b49-41d3-8490-ae2724128e67" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" exitCode=0 Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.372418 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" event={"ID":"7da49224-3b49-41d3-8490-ae2724128e67","Type":"ContainerDied","Data":"654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2"} Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.372767 5110 scope.go:117] "RemoveContainer" containerID="cb0b6120b2825b80ee94ecacb872639139788855e7496255ef0eb30ec1f502af" Mar 17 19:13:42 crc kubenswrapper[5110]: I0317 19:13:42.373440 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:13:42 crc kubenswrapper[5110]: E0317 19:13:42.373839 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.331869 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nl6jv/must-gather-s7gmb"] Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333082 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerName="smoketest-ceilometer" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333098 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerName="smoketest-ceilometer" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333113 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a207b29-155d-4d14-ba09-f2af999798e4" containerName="curl" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333118 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a207b29-155d-4d14-ba09-f2af999798e4" containerName="curl" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333145 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerName="smoketest-collectd" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333165 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerName="smoketest-collectd" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333275 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerName="smoketest-collectd" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333287 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="1a207b29-155d-4d14-ba09-f2af999798e4" containerName="curl" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.333296 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="58dbad41-8a71-475a-91cb-c18dd788dceb" containerName="smoketest-ceilometer" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.349689 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nl6jv/must-gather-s7gmb"] Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.349851 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.352280 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-nl6jv\"/\"default-dockercfg-ms2g5\"" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.352826 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-nl6jv\"/\"openshift-service-ca.crt\"" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.352973 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-nl6jv\"/\"kube-root-ca.crt\"" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.404518 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9kf5\" (UniqueName: \"kubernetes.io/projected/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-kube-api-access-f9kf5\") pod \"must-gather-s7gmb\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.404725 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-must-gather-output\") pod \"must-gather-s7gmb\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.506729 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f9kf5\" (UniqueName: \"kubernetes.io/projected/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-kube-api-access-f9kf5\") pod \"must-gather-s7gmb\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.506844 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-must-gather-output\") pod \"must-gather-s7gmb\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.507531 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-must-gather-output\") pod \"must-gather-s7gmb\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.542027 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9kf5\" (UniqueName: \"kubernetes.io/projected/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-kube-api-access-f9kf5\") pod \"must-gather-s7gmb\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:53 crc kubenswrapper[5110]: I0317 19:13:53.675920 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:13:54 crc kubenswrapper[5110]: I0317 19:13:54.114676 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nl6jv/must-gather-s7gmb"] Mar 17 19:13:54 crc kubenswrapper[5110]: I0317 19:13:54.471539 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" event={"ID":"e9446d25-1ae9-47db-b00f-a4c0bffe9d30","Type":"ContainerStarted","Data":"0037ed93694c094e1256bfe72428aa157c86afc5f27749a966ee399c106bc57e"} Mar 17 19:13:55 crc kubenswrapper[5110]: I0317 19:13:55.744563 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:13:55 crc kubenswrapper[5110]: E0317 19:13:55.744898 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.137233 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562914-ctfss"] Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.146370 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.150487 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.150869 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.150937 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.153248 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562914-ctfss"] Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.216818 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f86x\" (UniqueName: \"kubernetes.io/projected/c357e968-d23f-49e6-bf88-c7f214ce2bf7-kube-api-access-2f86x\") pod \"auto-csr-approver-29562914-ctfss\" (UID: \"c357e968-d23f-49e6-bf88-c7f214ce2bf7\") " pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.230422 5110 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.318789 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2f86x\" (UniqueName: \"kubernetes.io/projected/c357e968-d23f-49e6-bf88-c7f214ce2bf7-kube-api-access-2f86x\") pod \"auto-csr-approver-29562914-ctfss\" (UID: \"c357e968-d23f-49e6-bf88-c7f214ce2bf7\") " pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.346730 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f86x\" (UniqueName: \"kubernetes.io/projected/c357e968-d23f-49e6-bf88-c7f214ce2bf7-kube-api-access-2f86x\") pod \"auto-csr-approver-29562914-ctfss\" (UID: \"c357e968-d23f-49e6-bf88-c7f214ce2bf7\") " pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.490901 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.522154 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" event={"ID":"e9446d25-1ae9-47db-b00f-a4c0bffe9d30","Type":"ContainerStarted","Data":"1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874"} Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.522204 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" event={"ID":"e9446d25-1ae9-47db-b00f-a4c0bffe9d30","Type":"ContainerStarted","Data":"b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799"} Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.542287 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" podStartSLOduration=1.706192652 podStartE2EDuration="7.542267144s" podCreationTimestamp="2026-03-17 19:13:53 +0000 UTC" firstStartedPulling="2026-03-17 19:13:54.127605165 +0000 UTC m=+1578.160216687" lastFinishedPulling="2026-03-17 19:13:59.963679597 +0000 UTC m=+1583.996291179" observedRunningTime="2026-03-17 19:14:00.539509789 +0000 UTC m=+1584.572121331" watchObservedRunningTime="2026-03-17 19:14:00.542267144 +0000 UTC m=+1584.574878666" Mar 17 19:14:00 crc kubenswrapper[5110]: I0317 19:14:00.710459 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562914-ctfss"] Mar 17 19:14:01 crc kubenswrapper[5110]: I0317 19:14:01.527891 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562914-ctfss" event={"ID":"c357e968-d23f-49e6-bf88-c7f214ce2bf7","Type":"ContainerStarted","Data":"c6a9b6b615ab67e03df94be894c5074cd9fe4c02257480c77cf2ee3e37bd0bc8"} Mar 17 19:14:02 crc kubenswrapper[5110]: I0317 19:14:02.536409 5110 generic.go:358] "Generic (PLEG): container finished" podID="c357e968-d23f-49e6-bf88-c7f214ce2bf7" containerID="5084ea2de474eb12b5254b24b0f63cfdc8a778f9e0e996c94ebfc9eea120b2b0" exitCode=0 Mar 17 19:14:02 crc kubenswrapper[5110]: I0317 19:14:02.536480 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562914-ctfss" event={"ID":"c357e968-d23f-49e6-bf88-c7f214ce2bf7","Type":"ContainerDied","Data":"5084ea2de474eb12b5254b24b0f63cfdc8a778f9e0e996c94ebfc9eea120b2b0"} Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.040845 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.081761 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f86x\" (UniqueName: \"kubernetes.io/projected/c357e968-d23f-49e6-bf88-c7f214ce2bf7-kube-api-access-2f86x\") pod \"c357e968-d23f-49e6-bf88-c7f214ce2bf7\" (UID: \"c357e968-d23f-49e6-bf88-c7f214ce2bf7\") " Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.087824 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c357e968-d23f-49e6-bf88-c7f214ce2bf7-kube-api-access-2f86x" (OuterVolumeSpecName: "kube-api-access-2f86x") pod "c357e968-d23f-49e6-bf88-c7f214ce2bf7" (UID: "c357e968-d23f-49e6-bf88-c7f214ce2bf7"). InnerVolumeSpecName "kube-api-access-2f86x". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.182827 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2f86x\" (UniqueName: \"kubernetes.io/projected/c357e968-d23f-49e6-bf88-c7f214ce2bf7-kube-api-access-2f86x\") on node \"crc\" DevicePath \"\"" Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.791793 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562914-ctfss" Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.791800 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562914-ctfss" event={"ID":"c357e968-d23f-49e6-bf88-c7f214ce2bf7","Type":"ContainerDied","Data":"c6a9b6b615ab67e03df94be894c5074cd9fe4c02257480c77cf2ee3e37bd0bc8"} Mar 17 19:14:04 crc kubenswrapper[5110]: I0317 19:14:04.792268 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6a9b6b615ab67e03df94be894c5074cd9fe4c02257480c77cf2ee3e37bd0bc8" Mar 17 19:14:05 crc kubenswrapper[5110]: I0317 19:14:05.111716 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562908-xzlpd"] Mar 17 19:14:05 crc kubenswrapper[5110]: I0317 19:14:05.116010 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562908-xzlpd"] Mar 17 19:14:06 crc kubenswrapper[5110]: I0317 19:14:06.757647 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11779d6a-7707-4288-aaa3-8ae9411fb3e7" path="/var/lib/kubelet/pods/11779d6a-7707-4288-aaa3-8ae9411fb3e7/volumes" Mar 17 19:14:09 crc kubenswrapper[5110]: I0317 19:14:09.745007 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:14:09 crc kubenswrapper[5110]: E0317 19:14:09.745547 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:14:21 crc kubenswrapper[5110]: I0317 19:14:21.745092 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:14:21 crc kubenswrapper[5110]: E0317 19:14:21.745779 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:14:32 crc kubenswrapper[5110]: I0317 19:14:32.745904 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:14:32 crc kubenswrapper[5110]: E0317 19:14:32.746758 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:14:43 crc kubenswrapper[5110]: I0317 19:14:43.200733 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-75ffdb6fcd-xsmjk_57ae9010-5764-463a-93ad-d4036ee2b53c/control-plane-machine-set-operator/0.log" Mar 17 19:14:43 crc kubenswrapper[5110]: I0317 19:14:43.365386 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-755bb95488-xf26c_6ad554df-9a15-43bd-8691-7993f381e08c/kube-rbac-proxy/0.log" Mar 17 19:14:43 crc kubenswrapper[5110]: I0317 19:14:43.370448 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-755bb95488-xf26c_6ad554df-9a15-43bd-8691-7993f381e08c/machine-api-operator/0.log" Mar 17 19:14:46 crc kubenswrapper[5110]: I0317 19:14:46.757626 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:14:46 crc kubenswrapper[5110]: E0317 19:14:46.758605 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:14:49 crc kubenswrapper[5110]: I0317 19:14:49.005989 5110 scope.go:117] "RemoveContainer" containerID="007780b833e3d22bcc72a3bfa388d586fd85b933df1b8afe805cea9fc0ddf97b" Mar 17 19:14:56 crc kubenswrapper[5110]: I0317 19:14:56.020591 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-759f64656b-s8xn7_366966b1-9c92-430c-bc89-5a84ac8ed770/cert-manager-controller/0.log" Mar 17 19:14:56 crc kubenswrapper[5110]: I0317 19:14:56.126801 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-8966b78d4-5mcvs_0495d1f8-565a-45e1-ad95-c595cf5a894c/cert-manager-cainjector/0.log" Mar 17 19:14:56 crc kubenswrapper[5110]: I0317 19:14:56.198004 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-597b96b99b-nzg8p_88dd76ec-ef69-4d92-aaaf-bbe85e63b5c1/cert-manager-webhook/0.log" Mar 17 19:14:57 crc kubenswrapper[5110]: I0317 19:14:57.744660 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:14:57 crc kubenswrapper[5110]: E0317 19:14:57.744922 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.144336 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2"] Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.145328 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c357e968-d23f-49e6-bf88-c7f214ce2bf7" containerName="oc" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.145341 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="c357e968-d23f-49e6-bf88-c7f214ce2bf7" containerName="oc" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.145460 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="c357e968-d23f-49e6-bf88-c7f214ce2bf7" containerName="oc" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.166326 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2"] Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.166512 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.170289 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.170727 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.239402 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgmj2\" (UniqueName: \"kubernetes.io/projected/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-kube-api-access-lgmj2\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.239513 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-config-volume\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.239561 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-secret-volume\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.341086 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-secret-volume\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.341177 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lgmj2\" (UniqueName: \"kubernetes.io/projected/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-kube-api-access-lgmj2\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.341235 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-config-volume\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.342228 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-config-volume\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.351782 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-secret-volume\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.362224 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgmj2\" (UniqueName: \"kubernetes.io/projected/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-kube-api-access-lgmj2\") pod \"collect-profiles-29562915-msdc2\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.492154 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:00 crc kubenswrapper[5110]: I0317 19:15:00.916928 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2"] Mar 17 19:15:00 crc kubenswrapper[5110]: W0317 19:15:00.922544 5110 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76cb981d_46ee_41a9_b21e_bbd182b0e0e0.slice/crio-2d390d1128e854f289ed87e5e18bc9c7fc637615e5f620d383d175654d7d47b4 WatchSource:0}: Error finding container 2d390d1128e854f289ed87e5e18bc9c7fc637615e5f620d383d175654d7d47b4: Status 404 returned error can't find the container with id 2d390d1128e854f289ed87e5e18bc9c7fc637615e5f620d383d175654d7d47b4 Mar 17 19:15:01 crc kubenswrapper[5110]: I0317 19:15:01.193318 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" event={"ID":"76cb981d-46ee-41a9-b21e-bbd182b0e0e0","Type":"ContainerStarted","Data":"4880a417250e13b577d58a616534895f276067d5bb6ae6e8bcddf6143339e89b"} Mar 17 19:15:01 crc kubenswrapper[5110]: I0317 19:15:01.193361 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" event={"ID":"76cb981d-46ee-41a9-b21e-bbd182b0e0e0","Type":"ContainerStarted","Data":"2d390d1128e854f289ed87e5e18bc9c7fc637615e5f620d383d175654d7d47b4"} Mar 17 19:15:01 crc kubenswrapper[5110]: I0317 19:15:01.217933 5110 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" podStartSLOduration=1.217912539 podStartE2EDuration="1.217912539s" podCreationTimestamp="2026-03-17 19:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 19:15:01.212553453 +0000 UTC m=+1645.245164985" watchObservedRunningTime="2026-03-17 19:15:01.217912539 +0000 UTC m=+1645.250524061" Mar 17 19:15:02 crc kubenswrapper[5110]: I0317 19:15:02.205743 5110 generic.go:358] "Generic (PLEG): container finished" podID="76cb981d-46ee-41a9-b21e-bbd182b0e0e0" containerID="4880a417250e13b577d58a616534895f276067d5bb6ae6e8bcddf6143339e89b" exitCode=0 Mar 17 19:15:02 crc kubenswrapper[5110]: I0317 19:15:02.206398 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" event={"ID":"76cb981d-46ee-41a9-b21e-bbd182b0e0e0","Type":"ContainerDied","Data":"4880a417250e13b577d58a616534895f276067d5bb6ae6e8bcddf6143339e89b"} Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.454333 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.587745 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-config-volume\") pod \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.587940 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-secret-volume\") pod \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.588044 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgmj2\" (UniqueName: \"kubernetes.io/projected/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-kube-api-access-lgmj2\") pod \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\" (UID: \"76cb981d-46ee-41a9-b21e-bbd182b0e0e0\") " Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.588615 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-config-volume" (OuterVolumeSpecName: "config-volume") pod "76cb981d-46ee-41a9-b21e-bbd182b0e0e0" (UID: "76cb981d-46ee-41a9-b21e-bbd182b0e0e0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.595285 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "76cb981d-46ee-41a9-b21e-bbd182b0e0e0" (UID: "76cb981d-46ee-41a9-b21e-bbd182b0e0e0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.598232 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-kube-api-access-lgmj2" (OuterVolumeSpecName: "kube-api-access-lgmj2") pod "76cb981d-46ee-41a9-b21e-bbd182b0e0e0" (UID: "76cb981d-46ee-41a9-b21e-bbd182b0e0e0"). InnerVolumeSpecName "kube-api-access-lgmj2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.690211 5110 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.690277 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lgmj2\" (UniqueName: \"kubernetes.io/projected/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-kube-api-access-lgmj2\") on node \"crc\" DevicePath \"\"" Mar 17 19:15:03 crc kubenswrapper[5110]: I0317 19:15:03.690286 5110 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76cb981d-46ee-41a9-b21e-bbd182b0e0e0-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 19:15:04 crc kubenswrapper[5110]: I0317 19:15:04.221629 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" Mar 17 19:15:04 crc kubenswrapper[5110]: I0317 19:15:04.221649 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562915-msdc2" event={"ID":"76cb981d-46ee-41a9-b21e-bbd182b0e0e0","Type":"ContainerDied","Data":"2d390d1128e854f289ed87e5e18bc9c7fc637615e5f620d383d175654d7d47b4"} Mar 17 19:15:04 crc kubenswrapper[5110]: I0317 19:15:04.221721 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d390d1128e854f289ed87e5e18bc9c7fc637615e5f620d383d175654d7d47b4" Mar 17 19:15:09 crc kubenswrapper[5110]: I0317 19:15:09.860130 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-55568fc96c-6pbtm_8cb83be5-bf66-476c-b29e-f055352697e2/prometheus-operator/0.log" Mar 17 19:15:09 crc kubenswrapper[5110]: I0317 19:15:09.911552 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-646cb684f4-96h7l_ccfae76d-51c9-4f8c-a694-72d770811128/prometheus-operator-admission-webhook/0.log" Mar 17 19:15:09 crc kubenswrapper[5110]: I0317 19:15:09.989136 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx_391b404f-2c22-46cf-954e-84540c6fe899/prometheus-operator-admission-webhook/0.log" Mar 17 19:15:10 crc kubenswrapper[5110]: I0317 19:15:10.097961 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-587f9c8867-p6ncg_0a812097-b357-4b5e-bfa5-115b26efc4e3/operator/0.log" Mar 17 19:15:10 crc kubenswrapper[5110]: I0317 19:15:10.184222 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5b659947cf-krgf9_c9354dcb-0033-473e-a3f3-e01e7b35d9eb/perses-operator/0.log" Mar 17 19:15:11 crc kubenswrapper[5110]: I0317 19:15:11.745664 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:15:11 crc kubenswrapper[5110]: E0317 19:15:11.746164 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:15:22 crc kubenswrapper[5110]: I0317 19:15:22.745343 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:15:22 crc kubenswrapper[5110]: E0317 19:15:22.746344 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:15:23 crc kubenswrapper[5110]: I0317 19:15:23.808902 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/util/0.log" Mar 17 19:15:23 crc kubenswrapper[5110]: I0317 19:15:23.961360 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/pull/0.log" Mar 17 19:15:23 crc kubenswrapper[5110]: I0317 19:15:23.979975 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/util/0.log" Mar 17 19:15:23 crc kubenswrapper[5110]: I0317 19:15:23.996710 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/pull/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.138048 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/util/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.138842 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/pull/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.167372 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f48n9f_74637df8-ebd0-4afa-b472-12067b6b46c7/extract/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.286244 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/util/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.458103 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/pull/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.488848 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/util/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.497810 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/pull/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.636384 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/pull/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.654377 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/util/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.667829 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39erc4q8_8e42598c-1b8a-48bb-a6cb-58a5e0f703da/extract/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.782776 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/util/0.log" Mar 17 19:15:24 crc kubenswrapper[5110]: I0317 19:15:24.983934 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/util/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.006495 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/pull/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.042263 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/pull/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.157373 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/util/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.192372 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/extract/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.226503 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5rt6br_28ec609b-6d68-482d-8ce4-2a21d490f76a/pull/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.308938 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/util/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.453426 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/util/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.490349 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/pull/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.495747 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/pull/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.640870 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/util/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.672192 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/pull/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.712248 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726w9tdb_738d9d50-28e2-4982-bee6-7382cff35aaa/extract/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.841936 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/extract-utilities/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.961292 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/extract-utilities/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.984559 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/extract-content/0.log" Mar 17 19:15:25 crc kubenswrapper[5110]: I0317 19:15:25.985730 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/extract-content/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.170847 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/extract-content/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.177612 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/extract-utilities/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.267553 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qz95x_d9c6e43d-9340-4b90-879e-9508669131f1/registry-server/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.340302 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/extract-utilities/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.509238 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/extract-content/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.521933 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/extract-utilities/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.570577 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/extract-content/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.721647 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/extract-content/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.728619 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/extract-utilities/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.869170 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-547dbd544d-9m47z_53db1455-0000-4bb3-9f81-0e079a38a5ec/marketplace-operator/0.log" Mar 17 19:15:26 crc kubenswrapper[5110]: I0317 19:15:26.972739 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/extract-utilities/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.016086 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z69h6_53d18b2e-06d7-4ed4-bdb3-ce4766a804c1/registry-server/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.100617 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/extract-content/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.124858 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/extract-utilities/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.168600 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/extract-content/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.298498 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/extract-content/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.303460 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/extract-utilities/0.log" Mar 17 19:15:27 crc kubenswrapper[5110]: I0317 19:15:27.551037 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvkkc_293dea77-4f08-44cc-b93c-7eb90f01c833/registry-server/0.log" Mar 17 19:15:36 crc kubenswrapper[5110]: I0317 19:15:36.759446 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:15:36 crc kubenswrapper[5110]: E0317 19:15:36.760495 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:15:39 crc kubenswrapper[5110]: I0317 19:15:39.219391 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-55568fc96c-6pbtm_8cb83be5-bf66-476c-b29e-f055352697e2/prometheus-operator/0.log" Mar 17 19:15:39 crc kubenswrapper[5110]: I0317 19:15:39.255518 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-646cb684f4-96h7l_ccfae76d-51c9-4f8c-a694-72d770811128/prometheus-operator-admission-webhook/0.log" Mar 17 19:15:39 crc kubenswrapper[5110]: I0317 19:15:39.255623 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-646cb684f4-9z9tx_391b404f-2c22-46cf-954e-84540c6fe899/prometheus-operator-admission-webhook/0.log" Mar 17 19:15:39 crc kubenswrapper[5110]: I0317 19:15:39.346013 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-587f9c8867-p6ncg_0a812097-b357-4b5e-bfa5-115b26efc4e3/operator/0.log" Mar 17 19:15:39 crc kubenswrapper[5110]: I0317 19:15:39.384204 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5b659947cf-krgf9_c9354dcb-0033-473e-a3f3-e01e7b35d9eb/perses-operator/0.log" Mar 17 19:15:51 crc kubenswrapper[5110]: I0317 19:15:51.744599 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:15:51 crc kubenswrapper[5110]: E0317 19:15:51.745819 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.147709 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562916-9kvjh"] Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.150135 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="76cb981d-46ee-41a9-b21e-bbd182b0e0e0" containerName="collect-profiles" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.150264 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="76cb981d-46ee-41a9-b21e-bbd182b0e0e0" containerName="collect-profiles" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.150608 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="76cb981d-46ee-41a9-b21e-bbd182b0e0e0" containerName="collect-profiles" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.156404 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562916-9kvjh"] Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.157665 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.161663 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.162196 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.162451 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.284069 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5wsb\" (UniqueName: \"kubernetes.io/projected/5a8a45ff-e852-4cba-8218-fcfda160bcdb-kube-api-access-x5wsb\") pod \"auto-csr-approver-29562916-9kvjh\" (UID: \"5a8a45ff-e852-4cba-8218-fcfda160bcdb\") " pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.386606 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x5wsb\" (UniqueName: \"kubernetes.io/projected/5a8a45ff-e852-4cba-8218-fcfda160bcdb-kube-api-access-x5wsb\") pod \"auto-csr-approver-29562916-9kvjh\" (UID: \"5a8a45ff-e852-4cba-8218-fcfda160bcdb\") " pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.414281 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5wsb\" (UniqueName: \"kubernetes.io/projected/5a8a45ff-e852-4cba-8218-fcfda160bcdb-kube-api-access-x5wsb\") pod \"auto-csr-approver-29562916-9kvjh\" (UID: \"5a8a45ff-e852-4cba-8218-fcfda160bcdb\") " pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.482349 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:00 crc kubenswrapper[5110]: I0317 19:16:00.743205 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562916-9kvjh"] Mar 17 19:16:01 crc kubenswrapper[5110]: I0317 19:16:01.747269 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" event={"ID":"5a8a45ff-e852-4cba-8218-fcfda160bcdb","Type":"ContainerStarted","Data":"dd59616fe2be64e8ee67584bc4cdc385d76db69fa64d3c50c4be9d4a9cf4d2dc"} Mar 17 19:16:02 crc kubenswrapper[5110]: I0317 19:16:02.758331 5110 generic.go:358] "Generic (PLEG): container finished" podID="5a8a45ff-e852-4cba-8218-fcfda160bcdb" containerID="f31b5d8f1490f0d5e55754a821869790f4b00e4c03c4c5398cf82cd577a542e8" exitCode=0 Mar 17 19:16:02 crc kubenswrapper[5110]: I0317 19:16:02.758509 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" event={"ID":"5a8a45ff-e852-4cba-8218-fcfda160bcdb","Type":"ContainerDied","Data":"f31b5d8f1490f0d5e55754a821869790f4b00e4c03c4c5398cf82cd577a542e8"} Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.033521 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.142201 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5wsb\" (UniqueName: \"kubernetes.io/projected/5a8a45ff-e852-4cba-8218-fcfda160bcdb-kube-api-access-x5wsb\") pod \"5a8a45ff-e852-4cba-8218-fcfda160bcdb\" (UID: \"5a8a45ff-e852-4cba-8218-fcfda160bcdb\") " Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.166637 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a8a45ff-e852-4cba-8218-fcfda160bcdb-kube-api-access-x5wsb" (OuterVolumeSpecName: "kube-api-access-x5wsb") pod "5a8a45ff-e852-4cba-8218-fcfda160bcdb" (UID: "5a8a45ff-e852-4cba-8218-fcfda160bcdb"). InnerVolumeSpecName "kube-api-access-x5wsb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.243775 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-x5wsb\" (UniqueName: \"kubernetes.io/projected/5a8a45ff-e852-4cba-8218-fcfda160bcdb-kube-api-access-x5wsb\") on node \"crc\" DevicePath \"\"" Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.782331 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" event={"ID":"5a8a45ff-e852-4cba-8218-fcfda160bcdb","Type":"ContainerDied","Data":"dd59616fe2be64e8ee67584bc4cdc385d76db69fa64d3c50c4be9d4a9cf4d2dc"} Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.782439 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd59616fe2be64e8ee67584bc4cdc385d76db69fa64d3c50c4be9d4a9cf4d2dc" Mar 17 19:16:04 crc kubenswrapper[5110]: I0317 19:16:04.782478 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562916-9kvjh" Mar 17 19:16:05 crc kubenswrapper[5110]: I0317 19:16:05.124413 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562910-q8rvp"] Mar 17 19:16:05 crc kubenswrapper[5110]: I0317 19:16:05.137281 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562910-q8rvp"] Mar 17 19:16:05 crc kubenswrapper[5110]: I0317 19:16:05.746144 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:16:05 crc kubenswrapper[5110]: E0317 19:16:05.746654 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:16:06 crc kubenswrapper[5110]: I0317 19:16:06.760712 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0594a623-6ee6-4b42-beed-ef14ab5b2b72" path="/var/lib/kubelet/pods/0594a623-6ee6-4b42-beed-ef14ab5b2b72/volumes" Mar 17 19:16:16 crc kubenswrapper[5110]: I0317 19:16:16.746529 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:16:16 crc kubenswrapper[5110]: E0317 19:16:16.747581 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:16:20 crc kubenswrapper[5110]: E0317 19:16:20.414808 5110 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9446d25_1ae9_47db_b00f_a4c0bffe9d30.slice/crio-b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9446d25_1ae9_47db_b00f_a4c0bffe9d30.slice/crio-conmon-b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799.scope\": RecentStats: unable to find data in memory cache]" Mar 17 19:16:20 crc kubenswrapper[5110]: I0317 19:16:20.967532 5110 generic.go:358] "Generic (PLEG): container finished" podID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerID="b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799" exitCode=0 Mar 17 19:16:20 crc kubenswrapper[5110]: I0317 19:16:20.967629 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" event={"ID":"e9446d25-1ae9-47db-b00f-a4c0bffe9d30","Type":"ContainerDied","Data":"b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799"} Mar 17 19:16:20 crc kubenswrapper[5110]: I0317 19:16:20.968612 5110 scope.go:117] "RemoveContainer" containerID="b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799" Mar 17 19:16:21 crc kubenswrapper[5110]: I0317 19:16:21.300805 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nl6jv_must-gather-s7gmb_e9446d25-1ae9-47db-b00f-a4c0bffe9d30/gather/0.log" Mar 17 19:16:27 crc kubenswrapper[5110]: I0317 19:16:27.616123 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nl6jv/must-gather-s7gmb"] Mar 17 19:16:27 crc kubenswrapper[5110]: I0317 19:16:27.616837 5110 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="copy" containerID="cri-o://1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874" gracePeriod=2 Mar 17 19:16:27 crc kubenswrapper[5110]: I0317 19:16:27.619309 5110 status_manager.go:895] "Failed to get status for pod" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" err="pods \"must-gather-s7gmb\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-nl6jv\": no relationship found between node 'crc' and this object" Mar 17 19:16:27 crc kubenswrapper[5110]: I0317 19:16:27.626255 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nl6jv/must-gather-s7gmb"] Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.002383 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nl6jv_must-gather-s7gmb_e9446d25-1ae9-47db-b00f-a4c0bffe9d30/copy/0.log" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.003249 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.004479 5110 status_manager.go:895] "Failed to get status for pod" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" err="pods \"must-gather-s7gmb\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-nl6jv\": no relationship found between node 'crc' and this object" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.026668 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nl6jv_must-gather-s7gmb_e9446d25-1ae9-47db-b00f-a4c0bffe9d30/copy/0.log" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.026988 5110 generic.go:358] "Generic (PLEG): container finished" podID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerID="1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874" exitCode=143 Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.027140 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.027173 5110 scope.go:117] "RemoveContainer" containerID="1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.028610 5110 status_manager.go:895] "Failed to get status for pod" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" err="pods \"must-gather-s7gmb\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-nl6jv\": no relationship found between node 'crc' and this object" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.051231 5110 scope.go:117] "RemoveContainer" containerID="b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.068155 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-must-gather-output\") pod \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.068323 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9kf5\" (UniqueName: \"kubernetes.io/projected/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-kube-api-access-f9kf5\") pod \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\" (UID: \"e9446d25-1ae9-47db-b00f-a4c0bffe9d30\") " Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.075403 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-kube-api-access-f9kf5" (OuterVolumeSpecName: "kube-api-access-f9kf5") pod "e9446d25-1ae9-47db-b00f-a4c0bffe9d30" (UID: "e9446d25-1ae9-47db-b00f-a4c0bffe9d30"). InnerVolumeSpecName "kube-api-access-f9kf5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.118597 5110 scope.go:117] "RemoveContainer" containerID="1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874" Mar 17 19:16:28 crc kubenswrapper[5110]: E0317 19:16:28.119621 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874\": container with ID starting with 1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874 not found: ID does not exist" containerID="1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.119663 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874"} err="failed to get container status \"1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874\": rpc error: code = NotFound desc = could not find container \"1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874\": container with ID starting with 1a13d112ce3e66695a101a93930b086b6ad41d414120332e0f2b0245712a8874 not found: ID does not exist" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.119688 5110 scope.go:117] "RemoveContainer" containerID="b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799" Mar 17 19:16:28 crc kubenswrapper[5110]: E0317 19:16:28.121229 5110 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799\": container with ID starting with b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799 not found: ID does not exist" containerID="b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.121270 5110 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799"} err="failed to get container status \"b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799\": rpc error: code = NotFound desc = could not find container \"b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799\": container with ID starting with b381860d1bc880dee0e0586f904cfda4689a8eabf1dfad23acbf3a44bae00799 not found: ID does not exist" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.121555 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e9446d25-1ae9-47db-b00f-a4c0bffe9d30" (UID: "e9446d25-1ae9-47db-b00f-a4c0bffe9d30"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.170321 5110 reconciler_common.go:299] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.170373 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-f9kf5\" (UniqueName: \"kubernetes.io/projected/e9446d25-1ae9-47db-b00f-a4c0bffe9d30-kube-api-access-f9kf5\") on node \"crc\" DevicePath \"\"" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.359106 5110 status_manager.go:895] "Failed to get status for pod" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" pod="openshift-must-gather-nl6jv/must-gather-s7gmb" err="pods \"must-gather-s7gmb\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-nl6jv\": no relationship found between node 'crc' and this object" Mar 17 19:16:28 crc kubenswrapper[5110]: I0317 19:16:28.753662 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" path="/var/lib/kubelet/pods/e9446d25-1ae9-47db-b00f-a4c0bffe9d30/volumes" Mar 17 19:16:29 crc kubenswrapper[5110]: I0317 19:16:29.744745 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:16:29 crc kubenswrapper[5110]: E0317 19:16:29.745088 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:16:43 crc kubenswrapper[5110]: I0317 19:16:43.746051 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:16:43 crc kubenswrapper[5110]: E0317 19:16:43.747210 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:16:49 crc kubenswrapper[5110]: I0317 19:16:49.171453 5110 scope.go:117] "RemoveContainer" containerID="2ebdc7fb30206239dddc12c31f1ab0877d9e1a0f18ce9fc1dd435c91bcbd24ce" Mar 17 19:16:58 crc kubenswrapper[5110]: I0317 19:16:58.746233 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:16:58 crc kubenswrapper[5110]: E0317 19:16:58.747008 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:17:09 crc kubenswrapper[5110]: I0317 19:17:09.746477 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:17:09 crc kubenswrapper[5110]: E0317 19:17:09.747525 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:17:24 crc kubenswrapper[5110]: I0317 19:17:24.746185 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:17:24 crc kubenswrapper[5110]: E0317 19:17:24.747303 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:17:37 crc kubenswrapper[5110]: I0317 19:17:37.623223 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:17:37 crc kubenswrapper[5110]: I0317 19:17:37.628897 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:17:37 crc kubenswrapper[5110]: I0317 19:17:37.638770 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6kh2q_7346b312-05b5-4475-8d54-82709b69e7ed/kube-multus/0.log" Mar 17 19:17:37 crc kubenswrapper[5110]: I0317 19:17:37.643382 5110 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/1.log" Mar 17 19:17:39 crc kubenswrapper[5110]: I0317 19:17:39.745702 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:17:39 crc kubenswrapper[5110]: E0317 19:17:39.746302 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:17:51 crc kubenswrapper[5110]: I0317 19:17:51.745252 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:17:51 crc kubenswrapper[5110]: E0317 19:17:51.746351 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.144586 5110 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562918-cdlbt"] Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147100 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="gather" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147128 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="gather" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147152 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="copy" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147164 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="copy" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147189 5110 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5a8a45ff-e852-4cba-8218-fcfda160bcdb" containerName="oc" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147203 5110 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a8a45ff-e852-4cba-8218-fcfda160bcdb" containerName="oc" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147496 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="gather" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147536 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="e9446d25-1ae9-47db-b00f-a4c0bffe9d30" containerName="copy" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.147562 5110 memory_manager.go:356] "RemoveStaleState removing state" podUID="5a8a45ff-e852-4cba-8218-fcfda160bcdb" containerName="oc" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.152741 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.156325 5110 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-5gkpz\"" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.160863 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.164253 5110 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.168020 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562918-cdlbt"] Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.241870 5110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct58b\" (UniqueName: \"kubernetes.io/projected/8b89ec28-3425-43fe-8ce7-89b166da429b-kube-api-access-ct58b\") pod \"auto-csr-approver-29562918-cdlbt\" (UID: \"8b89ec28-3425-43fe-8ce7-89b166da429b\") " pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.343408 5110 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ct58b\" (UniqueName: \"kubernetes.io/projected/8b89ec28-3425-43fe-8ce7-89b166da429b-kube-api-access-ct58b\") pod \"auto-csr-approver-29562918-cdlbt\" (UID: \"8b89ec28-3425-43fe-8ce7-89b166da429b\") " pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.377469 5110 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct58b\" (UniqueName: \"kubernetes.io/projected/8b89ec28-3425-43fe-8ce7-89b166da429b-kube-api-access-ct58b\") pod \"auto-csr-approver-29562918-cdlbt\" (UID: \"8b89ec28-3425-43fe-8ce7-89b166da429b\") " pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.510364 5110 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.829948 5110 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562918-cdlbt"] Mar 17 19:18:00 crc kubenswrapper[5110]: I0317 19:18:00.882600 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" event={"ID":"8b89ec28-3425-43fe-8ce7-89b166da429b","Type":"ContainerStarted","Data":"a1b92eabcaf85010f7756b34b7c091aebd1bd0a203cc74ddb51575edad25d37f"} Mar 17 19:18:02 crc kubenswrapper[5110]: I0317 19:18:02.746312 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:18:02 crc kubenswrapper[5110]: E0317 19:18:02.748599 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:18:02 crc kubenswrapper[5110]: I0317 19:18:02.905093 5110 generic.go:358] "Generic (PLEG): container finished" podID="8b89ec28-3425-43fe-8ce7-89b166da429b" containerID="8b02758ec8443c46af26775e1fcc77ed7320ea293ade23c8ab2ef4eb37977f71" exitCode=0 Mar 17 19:18:02 crc kubenswrapper[5110]: I0317 19:18:02.905609 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" event={"ID":"8b89ec28-3425-43fe-8ce7-89b166da429b","Type":"ContainerDied","Data":"8b02758ec8443c46af26775e1fcc77ed7320ea293ade23c8ab2ef4eb37977f71"} Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.250014 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.300557 5110 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct58b\" (UniqueName: \"kubernetes.io/projected/8b89ec28-3425-43fe-8ce7-89b166da429b-kube-api-access-ct58b\") pod \"8b89ec28-3425-43fe-8ce7-89b166da429b\" (UID: \"8b89ec28-3425-43fe-8ce7-89b166da429b\") " Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.307826 5110 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b89ec28-3425-43fe-8ce7-89b166da429b-kube-api-access-ct58b" (OuterVolumeSpecName: "kube-api-access-ct58b") pod "8b89ec28-3425-43fe-8ce7-89b166da429b" (UID: "8b89ec28-3425-43fe-8ce7-89b166da429b"). InnerVolumeSpecName "kube-api-access-ct58b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.403382 5110 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ct58b\" (UniqueName: \"kubernetes.io/projected/8b89ec28-3425-43fe-8ce7-89b166da429b-kube-api-access-ct58b\") on node \"crc\" DevicePath \"\"" Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.925788 5110 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" event={"ID":"8b89ec28-3425-43fe-8ce7-89b166da429b","Type":"ContainerDied","Data":"a1b92eabcaf85010f7756b34b7c091aebd1bd0a203cc74ddb51575edad25d37f"} Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.926233 5110 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1b92eabcaf85010f7756b34b7c091aebd1bd0a203cc74ddb51575edad25d37f" Mar 17 19:18:04 crc kubenswrapper[5110]: I0317 19:18:04.925851 5110 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562918-cdlbt" Mar 17 19:18:05 crc kubenswrapper[5110]: I0317 19:18:05.341979 5110 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562912-mvgwc"] Mar 17 19:18:05 crc kubenswrapper[5110]: I0317 19:18:05.352403 5110 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562912-mvgwc"] Mar 17 19:18:06 crc kubenswrapper[5110]: I0317 19:18:06.756769 5110 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b686fe0d-dc7d-4a76-82c4-2ee230ead1f4" path="/var/lib/kubelet/pods/b686fe0d-dc7d-4a76-82c4-2ee230ead1f4/volumes" Mar 17 19:18:16 crc kubenswrapper[5110]: I0317 19:18:16.763666 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:18:16 crc kubenswrapper[5110]: E0317 19:18:16.764550 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" Mar 17 19:18:31 crc kubenswrapper[5110]: I0317 19:18:31.745491 5110 scope.go:117] "RemoveContainer" containerID="654fef00ba808e4d283f3c2f13fd8360e228504c8373323370b0f0d9ff7cfdb2" Mar 17 19:18:31 crc kubenswrapper[5110]: E0317 19:18:31.746561 5110 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6jbgs_openshift-machine-config-operator(7da49224-3b49-41d3-8490-ae2724128e67)\"" pod="openshift-machine-config-operator/machine-config-daemon-6jbgs" podUID="7da49224-3b49-41d3-8490-ae2724128e67" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515156324423024452 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015156324423017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015156320275016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015156320275015463 5ustar corecore